678 lines
28 KiB
Plaintext
678 lines
28 KiB
Plaintext
--------------------------------------------------------------------
|
||
README.TXT
|
||
|
||
Adaptec Storage Manager (ASM)
|
||
|
||
as of May 7, 2012
|
||
--------------------------------------------------------------------
|
||
Please review this file for important information about issues
|
||
and erratas that were discovered after completion of the standard
|
||
product documentation. In the case of conflict between various
|
||
parts of the documentation set, this file contains the most
|
||
current information.
|
||
|
||
The following information is available in this file:
|
||
|
||
1. Software Versions and Documentation
|
||
1.1 Adaptec Storage Manager
|
||
1.2 Documentation
|
||
2. Installation and Setup Notes
|
||
2.1 Supported Operating Systems
|
||
2.2 Minimum System Requirements
|
||
2.3 General Setup Notes
|
||
2.4 Linux Setup Notes
|
||
2.5 Debian Linux Setup Notes
|
||
3. General Cautions and Notes
|
||
3.1 General Cautions
|
||
3.2 General Notes
|
||
4. Operating System-Specific Issues and Notes
|
||
4.1 Windows - All
|
||
4.2 Windows 64-Bit
|
||
4.3 Linux
|
||
4.4 Debian and Ubuntu
|
||
4.5 FreeBSD
|
||
4.6 Fedora and FreeBSD
|
||
4.7 Linux and FreeBSD
|
||
4.8 VMware
|
||
5. RAID Level-Specific Notes
|
||
5.1 RAID 1 and RAID 5 Notes
|
||
5.2 RAID 10 Notes
|
||
5.3 RAID x0 Notes
|
||
5.4 RAID Volume Notes
|
||
5.5 JBOD Notes
|
||
5.6 Hybrid RAID Notes
|
||
5.7 RAID-Level Migration Notes
|
||
6. Power Management Issues and Notes
|
||
7. "Call Home" Issues and Notes
|
||
8. ARCCONF Issues and Notes
|
||
9. Other Issues and Notes
|
||
|
||
--------------------------------------------------------------------
|
||
1. Software Versions and Documentation
|
||
|
||
1.1. Adaptec Storage Manager Version 7.3.1, ARCCONF Version 7.3.1
|
||
|
||
1.2. Documentation on this DVD
|
||
|
||
PDF format*:
|
||
|
||
- Adaptec Storage Manager User's Guide
|
||
- Adaptec RAID Controller Command Line Utility User's Guide
|
||
|
||
*Requires Adobe Acrobat Reader 4.0 or later
|
||
|
||
HTML and text format:
|
||
|
||
- Adaptec Storage Manager Online Help
|
||
- Adaptec Storage Manager README.TXT file
|
||
|
||
--------------------------------------------------------------------
|
||
2. Installation and Setup Notes
|
||
|
||
- The Adaptec Storage Manager User's Guide contains complete installation
|
||
instructions for the Adaptec Storage Manager software. The Adaptec
|
||
RAID Controllers Command Line Utility User's Guide contains
|
||
complete installation instructions for ARCCONF, Remote ARCCONF,
|
||
and the Adaptec CIM Provider. The Adaptec RAID Controllers
|
||
Installation and User's Guide contains complete installation
|
||
instructions for Adaptec RAID controllers and drivers.
|
||
|
||
2.1 Supported Operating Systems
|
||
|
||
- Microsoft Windows*:
|
||
|
||
o Windows Server 2008, 32-bit and 64-bit
|
||
o Windows Server 2008 R2, 64-bit
|
||
o Windows SBS 2011, 32-bit and 64-bit
|
||
o Windows Storage Server 2008 R2, 32-bit and 64-bit
|
||
o Windows Storage Server 2011, 32-bit and 64-bit
|
||
o Windows 7, 32-bit and 64-bit
|
||
|
||
*Out-of-box and current service pack
|
||
|
||
- Linux:
|
||
|
||
o Red Hat Enterprise Linux 5.7, 6.1, IA-32 and x64
|
||
o SuSE Linux Enterprise Server 10, 11, IA-32 and x64
|
||
o Debian Linux 5.0.7, 6.0 IA-32 and x64
|
||
o Ubuntu Linux 10.10, 11.10, IA-32 and x64
|
||
o Fedora Linux 14, 15, 16 IA-32 and x64
|
||
o Cent OS 5.7, 6.2, IA-32 and x64
|
||
o VMware ESXi 5.0, VMware ESX 4.1 Classic (Agent only)
|
||
|
||
- Solaris:
|
||
|
||
o Solaris 10,
|
||
o Solaris 11 Express
|
||
|
||
- FreeBSD:
|
||
|
||
o FreeBSD 7.4, 8.2
|
||
|
||
2.2 Minimum System Requirements
|
||
|
||
o Pentium Compatible 1.2 GHz processor, or equivalent
|
||
o 512 MB RAM
|
||
o 135 MB hard disk drive space
|
||
o Greater than 256 color video mode
|
||
|
||
2.3 General Setup Notes
|
||
|
||
- You can configure Adaptec Storage Manager settings on other
|
||
servers exactly as they are configured on one server. To
|
||
replicate the Adaptec Storage Manager Enterprise view tree
|
||
and notification list, do the following:
|
||
|
||
1. Install Adaptec Storage Manager on one server.
|
||
|
||
2. Start Adaptec Storage Manager. Using the 'Add remote system'
|
||
action, define the servers for your tree.
|
||
|
||
3. Open the Notification Manager. Using the 'Add system'
|
||
action, define the notification list.
|
||
|
||
4. Exit Adaptec Storage Manager.
|
||
|
||
5. Copy the following files onto a diskette from the directory
|
||
where the Adaptec Storage Manager is installed:
|
||
|
||
RaidMSys.ser --> to replicate the tree
|
||
RaidNLst.ser --> to replicate the notification list
|
||
RaidSMTP.ser --> to replicate the SMTP e-mail notification list
|
||
RaidJob.ser --> to replicate the jobs in the Task Scheduler
|
||
|
||
6. Install Adaptec Storage Manager on the other servers.
|
||
|
||
7. Copy the files from the diskette into the directory where
|
||
Adaptec Storage Manager is installed on the other servers.
|
||
|
||
2.4 Linux Setup Notes
|
||
|
||
- Because the RPM for Red Hat Enterprise Linux 5 is unsigned, the
|
||
installer reports that the package is "Unsigned, Malicious Software".
|
||
Ignore the message and continue the installation.
|
||
|
||
- To run Adaptec Storage Manager under Red Hat Enterprise Linux for
|
||
x64, the Standard installation with "Compatibility Arch Support"
|
||
is required.
|
||
|
||
- To install Adaptec Storage Manager on Red Hat Enterprise Linux,
|
||
you must install two packages from the Red Hat installation CD:
|
||
|
||
o compat-libstdc++-7.3-2.96.122.i386.rpm
|
||
o compat-libstdc++--devel-7.3-2.96.122.i386.rpm
|
||
|
||
NOTE: The version string in the file name may be different
|
||
from above. Be sure to check the version string on the
|
||
Red Hat CD.
|
||
|
||
For example, type:
|
||
|
||
rpm --install /mnt/compat-libstdc++-7.3-2.96.122.i386.rpm
|
||
|
||
where mnt is the mount point of the CD-ROM drive.
|
||
|
||
- To install Adaptec Storage Manager on Red Hat Enterprise Linux 5,
|
||
you must install one of these packages from the Red Hat
|
||
installation CD:
|
||
|
||
o libXp-1.0.0-8.i386.rpm (32-Bit)
|
||
o libXp-1.0.0-8.x86.rpm (64-Bit)
|
||
|
||
- To install Adaptec Storage Manager on SuSE Linux Enterprise
|
||
Desktop 9, Service Pack 1, for 64-bit systems, you must install
|
||
two packages from the SuSE Linux installation CD:
|
||
|
||
- liblcms-devel-1.12-55.2.x86_64.rpm
|
||
- compat-32bit-9-200502081830.x86_64.rpm
|
||
|
||
NOTE: The version string in the file name may be different
|
||
from above. Be sure to check the version string on the
|
||
installation CD.
|
||
|
||
- To enable ASM's hard drive firmware update feature on RHEL 64-bit
|
||
systems, you must ensure that the "sg" module is loaded in the
|
||
kernel. To load the module manually (if it is not loaded already),
|
||
use the command "modprobe sg".
|
||
|
||
2.5 Debian Linux Setup Notes
|
||
|
||
- You can use the ASM GUI on Debian Linux 5.x only if you install
|
||
the GNOME desktop. Due to a compatibility issue with X11, the
|
||
default KDE desktop is not supported in this release.
|
||
|
||
- To ensure that the ASM Agent starts automatically when Debian
|
||
is rebooted, you must update the default start and stop values
|
||
in /etc/init.d/stor_agent, as follows:
|
||
|
||
<20>[Original]
|
||
# Default-Start: 2 3 5
|
||
# Default-Stop: 0 1 2 6
|
||
|
||
<20>[Modification]
|
||
# Default-Start: 2 3 4 5
|
||
# Default-Stop: 0 1 6
|
||
|
||
To activate the changes, execute 'insserv stor_agent', as root.
|
||
|
||
--------------------------------------------------------------------
|
||
3. Adaptec Storage Manager General Cautions and Notes
|
||
|
||
3.1 General Cautions
|
||
|
||
- This release supports a maximum of 8 concurrent online capacity
|
||
expansion (OCE) tasks in the RAID array migration wizard.
|
||
|
||
- While building or clearing a logical drive, do not remove and
|
||
re-insert any drive from that logical drive. Doing so may cause
|
||
unpredictable results.
|
||
|
||
- Do not move disks comprising a logical drive from one controller
|
||
to another while the power is on. Doing so could cause the loss of
|
||
the logical drive configuration or data, or both. Instead, power
|
||
off both affected controllers, move the drives, and then restart.
|
||
|
||
- When using Adaptec Storage Manager and the CLI concurrently,
|
||
configuration changes may not appear in the Adaptec Storage
|
||
Manager GUI until you refresh the display (by pressing F5).
|
||
|
||
3.2 General Notes
|
||
|
||
- Adaptec Storage Manager requires the following range of ports
|
||
to be open for remote access: 34570-34580 (TCP), 34570 (UDP),
|
||
34577-34580 (UDP).
|
||
|
||
- Adaptec Storage Manager generates log files automatically to
|
||
assist in tracking system activity. The log files are
|
||
created in the directory where Adaptec Storage Manager is
|
||
installed.
|
||
|
||
o RaidEvt.log - Contains the information reported in
|
||
Adaptec Storage Manager event viewer for all
|
||
local and remote systems.
|
||
|
||
o RaidEvtA.log - Contains the information reported in
|
||
Adaptec Storage Manager event viewer for the
|
||
local system.
|
||
|
||
o RaidNot.log - Contains the information reported in the
|
||
Notification Manager event viewer.
|
||
|
||
o RaidErr.log - Contains Java messages generated by
|
||
Adaptec Storage Manager.
|
||
|
||
o RaidErrA.log - Contains Java messages generated by the
|
||
Adaptec Storage Manager agent.
|
||
|
||
o RaidCall.log - Contains the information reported when
|
||
statistics logging is enabled in ASM.
|
||
|
||
Information written to these files is appended to the existing
|
||
files to maintain a history. However, when an error log file
|
||
reaches a size of 5 Mbytes, it is copied to a new file with
|
||
the extension .1 and the original (that is, the .LOG file) is
|
||
deleted and recreated. For other log files, a .1 file is created
|
||
when the .LOG file reaches a size of 1 Mbyte. If a .1 file already
|
||
exists, the existing .1 file is destroyed.
|
||
|
||
- In the Event viewer, Adaptec Storage Manager reports both the
|
||
initial build task for a logical drive and a subsequent Verify/Fix
|
||
as a "Build/Verify" task.
|
||
|
||
- When displaying information about a physical device, the device,
|
||
vendor and model information may be displayed incorrectly.
|
||
|
||
- After using a hot spare to successfully rebuild a redundant
|
||
logical drive, Adaptec Storage Manager will continue to
|
||
show the drive as a global hot spare. To remove the hot spare
|
||
designation, delete it in Adaptec Storage Manager.
|
||
|
||
--------------------------------------------------------------------
|
||
4. Operating System-Specific Issues and Notes
|
||
|
||
4.1 Windows - All
|
||
|
||
- The Java Virtual Machine has a problem with the 256-color
|
||
palette. (The Adaptec Storage Manager display may be distorted
|
||
or hard to read.) Set the Display Properties Settings to a
|
||
color mode with greater than 256 colors.
|
||
|
||
- When you shut down Windows, you might see the message
|
||
"unexpected shutdown". Windows displays this message if the
|
||
Adaptec Storage Manager Agent fails to exit within 3 seconds.
|
||
It has no effect on file I/O or other system operations and can
|
||
be ignored.
|
||
|
||
4.2 Windows 64-Bit
|
||
|
||
- Adaptec RAID controllers do not produce an audible alarm on the
|
||
following 64-bit Windows operating systems:
|
||
|
||
o Windows Server 2003 x64 Edition (all versions)
|
||
|
||
4.3 Linux
|
||
|
||
- When you delete a logical drive, the operating system can no longer
|
||
see the last logical drive. WORKAROUND: To allow Linux to see the
|
||
last logical drive, restart your system.
|
||
|
||
- The controller does not support attached CD drives during OS
|
||
installation.
|
||
|
||
- On certain versions of Linux, you may see messages concerning font
|
||
conversion errors. Font configuration under X-Windows is a known
|
||
JVM problem. It does not affect the proper operation of the
|
||
Adaptec Storage Manager software. To suppress these messages,
|
||
add the following line to your .Xdefaults file:
|
||
|
||
stringConversionWarnings: False
|
||
|
||
4.4 Debian and Ubuntu
|
||
|
||
- To create logical drives on Debian and Ubuntu installations, you
|
||
must log in as root. It is not sufficient to start ASM with the
|
||
'sudo /usr/StorMan/StorMan.sh' command (when not logged in as
|
||
root). WORKAROUND: To create logical drives on Ubuntu when not
|
||
logged in as root, install the package 'sudo dpkg -i storm_6.50-15645_amd64.deb'.
|
||
|
||
4.5 FreeBSD
|
||
|
||
- On FreeBSD systems, JBOD disks created with Adaptec Storage Manager
|
||
are not immediately available to the OS. You must reboot the
|
||
system before you can use the JBOD.
|
||
|
||
4.6 Fedora and FreeBSD
|
||
|
||
- Due to an issue with the Java JDialog Swing class, the 'Close'
|
||
button may not appear on some Adaptec Storage Manager windows
|
||
or dialog boxes under FreeBSD or Fedora Linux 15 or higher.
|
||
WORKAROUND: Press ALT+F4 or right-click on the title bar, then
|
||
close the dialog box from the pop-up menu.
|
||
|
||
4.7 Linux and FreeBSD
|
||
|
||
- If you cannot connect to a local or remote Adaptec Storage Manager
|
||
installed on a Linux or FreeBSD system, verify that the TCP/IP hosts
|
||
file is configured properly.
|
||
|
||
1. Open the /etc/hosts file.
|
||
|
||
NOTE: The following is an example:
|
||
|
||
127.0.0.1 localhost.localdomain localhost matrix
|
||
|
||
2. If the hostname of the system is identified on the line
|
||
with 127.0.0.1, you must create a new host line.
|
||
|
||
3. Remove the hostname from the 127.0.0.1 line.
|
||
|
||
NOTE: The following is an example:
|
||
|
||
127.0.0.1 localhost.localdomain localhost
|
||
|
||
4. On a new line, type the IP address of the system.
|
||
|
||
5. Using the Tab key, tab to the second column and enter the
|
||
fully qualified hostname.
|
||
|
||
6. Using the Tab key, tab to the third column and enter the
|
||
nickname for the system.
|
||
|
||
NOTE: The following is an example of a completed line:
|
||
|
||
1.1.1.1 matrix.localdomain matrix
|
||
|
||
where 1.1.1.1 is the IP address of the server and
|
||
matrix is the hostname of the server.
|
||
|
||
7. Restart the server for the changes to take effect.
|
||
|
||
4.8 VMware
|
||
|
||
- If you are unable to connect to VMware ESX Server from a
|
||
remote ASM GUI, even though it appears in the Enterprise
|
||
View as a remote system, most likely, some required ports
|
||
are open and others are not. (The VMware ESX firewall blocks
|
||
most ports, by default.) Check to make sure that all ports
|
||
34570 thru 34581 are opened on the ESX server.
|
||
|
||
- After making array configuration changes in VMware, you must
|
||
run the "esxcfg-rescan" tool manually at the VMware console
|
||
to notify the operating system of the new target characteristics
|
||
and/or availability. Alternatively, you can rescan from the
|
||
Virtual Infrastructure Client: click on the host in the left
|
||
panel, select the Configuration tab, choose "Storage Adapters",
|
||
then, on the right side of the screen, click "Rescan".
|
||
|
||
- With VMware ESX 4.1, the OS command 'esxcfg-scsidevs -a'
|
||
incorrectly identifies the Adaptec ASR-5445 controller as
|
||
"Adaptec ASR5800". (ASM itself identifies the controller
|
||
correctly.) To verify the controller name at the OS level,
|
||
use this command to check the /proc file system:
|
||
|
||
# cat /proc/scsi/aacraid/<Node #>
|
||
|
||
where <Node #> is 1, 2, 3 etc.
|
||
|
||
--------------------------------------------------------------------
|
||
5. RAID Level-Specific Notes
|
||
|
||
5.1 RAID 1 and RAID 5 Notes
|
||
|
||
- During a logical device migration from RAID 1 or RAID 5 to
|
||
RAID 0, if the original logical drive had a spare drive
|
||
attached, the resulting RAID 0 retains the spare drive.
|
||
Since RAID 0 is not redundant, you can remove the hot spare.
|
||
|
||
5.2 RAID 10 Notes
|
||
|
||
- If you force online a failed RAID 10, ASM erroneously shows two
|
||
drives rebuilding (the two underlying member drives), not one.
|
||
|
||
- You cannot change the priority of a RAID 10 verify. Setting
|
||
the priority at the start of a verify has no effect. The
|
||
priority is still shown as high. Changing the priority of
|
||
a running verify on a RAID-10 changes the displayed priority
|
||
until a rescan is done, then the priority shows as high again.
|
||
|
||
- Performing a Verify or Verify/Fix on an RAID 10 displays the
|
||
same message text in the event log: "Build/Verify started on
|
||
second level logical drive of 'LogicalDrive_0.'" You may see the
|
||
message three times for a Verify, but only once for a Verify/Fix.
|
||
|
||
5.3 RAID x0 Notes
|
||
|
||
- To create a RAID x0 with an odd number of drives (15, 25, etc),
|
||
specify an odd number of second-level devices in the Advanced
|
||
settings for the array. For a 25 drive RAID 50, for instance,
|
||
the default is 24 drives.
|
||
|
||
NOTE: This differs from the BIOS utility, which creates RAID x0
|
||
arrays with an odd number of drives by default.
|
||
|
||
- After building or verifying a leg of a second-level logical drive,
|
||
the status of the second-level logical drive is displayed as a
|
||
"Quick Initialized" drive.
|
||
|
||
5.4 RAID Volume Notes
|
||
|
||
- In ASM, a failed RAID Volume comprised of two RAID 1 logical
|
||
drives is erroneously reported as a failed RAID 10. A failed
|
||
RAID Volume comprised of two RAID 5 logical drives is
|
||
erroneously reported as a failed RAID 50.
|
||
|
||
5.5 JBOD Notes
|
||
|
||
- In this release, ASM deletes partitioned JBODs without issuing
|
||
a warning message.
|
||
|
||
- When migrating a JBOD to a Simple Volume, the disk must be quiescent
|
||
(no I/O load). Otherwise, the migration will fail with an I/O Read
|
||
error.
|
||
|
||
5.6 Hybrid RAID Notes
|
||
|
||
- ASM supports Hybrid RAID 1 and RAID 10 logical drives comprised
|
||
of hard disk drives and Solid State Drives (SSDs). For a Hybrid
|
||
RAID 10, you must select an equal number of SSDs and HDDs in
|
||
<20>every other drive<76> order, that is: SSD<53>HDD<44>SSD<53>HDD, and so on.
|
||
Failure to select drives in this order creates a standard
|
||
logical drive that does not take advantage of SSD performance.
|
||
|
||
5.7 RAID-Level Migration (RLM) Notes
|
||
|
||
- We strongly recommend that you use the default 256KB stripe
|
||
size for all RAID-level migrations. Choosing a different stripe
|
||
size may crash the system.
|
||
|
||
- If a disk error occurs when migrating a 2TB RAID 0 to RAID 5
|
||
(eg, bad blocks), ASM displays a message that the RAID 5 drive
|
||
is reconfiguring even though the migration failed and no
|
||
RAID-level migration task is running. To recreate the
|
||
logical drive, fix or replace the bad disk, delete the RAID 5
|
||
in ASM, then try again.
|
||
|
||
- When migrating a RAID 5EE, be careful not to remove and re-insert
|
||
a drive in the array. If you do, the drive will not be included
|
||
when the array is rebuilt. The migration will stop and the drive
|
||
will be reported as Ready (not part of array).
|
||
|
||
NOTE: We strongly recommend that you do not remove and re-insert
|
||
any drive during a RAID-level migration.
|
||
|
||
- When migrating a RAID 6 to a RAID 5, the migration will fail if
|
||
the (physical) drive order on the target logical device differs
|
||
from the source; for instance, migrating a four-drive RAID 6 to
|
||
a three-drive RAID 5.
|
||
|
||
- Migrating a RAID 5 with greater than 2TB capacity to RAID 6 or
|
||
RAID 10 is not supported in this release. Doing so may crash
|
||
the system.
|
||
|
||
- When migrating from a RAID 0 to any redundant logical drive,
|
||
like RAID 5 or 10, Adaptec Storage Manager shows the status as
|
||
"Degraded Reconfiguring" for a moment, then the status changes
|
||
to "Reconfiguring." The "Degraded" status does not appear in
|
||
the event log.
|
||
|
||
- The following RAID-level migrations and online capacity
|
||
expansions (OCE) are NOT supported:
|
||
|
||
o RAID 50 to RAID 5 RLM
|
||
o RAID 60 to RAID 6 RLM
|
||
o RAID 50 to RAID 60 OCE
|
||
|
||
- During a RAID-level migration, ASM and the BIOS utility show
|
||
different RAID levels while the migration is in progress. ASM shows
|
||
the target RAID level; the BIOS utility shows the current RAID level.
|
||
|
||
- If a disk error occurs during a RAID-level migration (eg, bad blocks),
|
||
the exception is reported in the ASM event viewer (bottom pane)
|
||
and in the support archive file (Support.zip, Controller 1 logs.txt),
|
||
but not in the main ASM Event Log file, RaidEvtA.log.
|
||
|
||
- Always allow a RAID-level migration to complete before gathering
|
||
support archive information in Support.zip. Otherwise, the Support.zip
|
||
file will include incorrect partition information. Once the RLM is
|
||
complete, the partition information will be reported correctly.
|
||
|
||
--------------------------------------------------------------------
|
||
6. Power Management Issues and Notes
|
||
|
||
- You must use a compatible combination of Adaptec Storage Manager
|
||
and controller firmware and driver software to use the power
|
||
management feature. All software components must support power
|
||
management. You can download the latest controller firmware
|
||
and drivers from the Adaptec Web site at www.adaptec.com.
|
||
|
||
- Power management is not supported under FreeBSD.
|
||
|
||
- Power management settings apply only to logical drives in the
|
||
Optimal state. If you change the power settings on a Failed
|
||
logical drive, then force the drive online, the previous
|
||
settings are reinstated.
|
||
|
||
- After setting power values for a logical drive in ARCCONF, the
|
||
settings are not updated in the Adaptec Storage Manager GUI.
|
||
|
||
--------------------------------------------------------------------
|
||
7. "Call Home" Issues and Notes
|
||
|
||
- The Call Home feature is not supported in this release. To gather
|
||
statistics about your system for remote analysis, enable statistics
|
||
logging in ASM, then create a Call Home Support Archive. For more
|
||
information, see the user's guide.
|
||
|
||
--------------------------------------------------------------------
|
||
8. ARCCONF Issues and Notes
|
||
|
||
- With VMware ESX 4.1, you cannot delete a logical drive
|
||
with ARCCONF. WORKAROUND: Connect to the VMware machine from a
|
||
remote ASM GUI, then delete the logical drive.
|
||
|
||
- With Linux kernel versions 2.4 and 2.6, the ARCCONF
|
||
DELETE <logical_drive> command may fail with a Kernel Oops
|
||
error message. Even though the drives are removed from the
|
||
Adaptec Storage Manager GUI, they may not really be deleted.
|
||
Reboot the controller; then, issue the ARCCONF DELETE command
|
||
again.
|
||
|
||
--------------------------------------------------------------------
|
||
9. Other Issues and Notes
|
||
|
||
- Some solid state drives identify themselves as ROTATING media.
|
||
As a result, these SSDs:
|
||
|
||
o Appear as SATA drives in the ASM Physical Devices View
|
||
o Cannot be used as Adaptec maxCache devices
|
||
o Cannot be used within a hybrid RAID array (comprised of
|
||
SSDs and hard disks)
|
||
|
||
- The blink pattern on Adaptec Series 6/6Q/6E/6T controllers differs
|
||
from Series 2 and Series 5 controllers:
|
||
|
||
o When blinking drives in ASM, the red LED goes on and stays solid;
|
||
on Series 2 and 5 controllers, the LED blinks on and off.
|
||
|
||
o When failing drives in ASM (using the 'Set drive state to failed'
|
||
action), the LED remains off; on Series 2 and 5 controllers, the
|
||
LED goes on and remains solid.
|
||
|
||
- Cache settings for RAID Volumes (Read cache, Write cache, maxCache)
|
||
have no effect. The cache settings for the underlying logical
|
||
devices take priority.
|
||
|
||
- On rare occasions, ASM will report invalid medium error counts on
|
||
a SATA hard drive or SSD. To correct the problem, use ARCCONF to
|
||
clear the device counts. The command is:
|
||
|
||
arcconf getlogs <Controller_ID> DEVICE clear
|
||
|
||
- On rare occasions, ASM lists direct-attached hard drives and SSDs
|
||
as drives in a virtual SGPIO enclosure. Normally, the drives are
|
||
listed in the Physical Devices View under ports CN0 and CN1.
|
||
|
||
- Hard Drive Firmware Update Wizard:
|
||
|
||
o Firmware upgrade on Western Digital WD5002ABYS-01B1B0 hard drives
|
||
is not supported for packet sizes below 2K (512/1024).
|
||
|
||
o After flashing the firmware of a Seagate Barracuda ES ST3750640NS
|
||
hard drive, you MUST cycle the power before ASM will show the new
|
||
image. You can pull out and re-insert the drive; power cycle the
|
||
enclosure; or power cycle the system if the drive is attached directly.
|
||
|
||
- Secure Erase:
|
||
|
||
o If you reboot the system while a Secure Erase operation is in
|
||
progress, the affected drive may not be displayed in Adaptec
|
||
Storage Manager or other Adaptec utilities, such as the ACU.
|
||
|
||
o You can perform a Secure Erase on a Solid State Drive (SSD) to
|
||
remove the metadata. However, the drive will move to the Failed
|
||
state when you reboot the system. To use the SSD, reboot to
|
||
the BIOS, then initialize the SSD. After initialization, the SSD
|
||
will return to the Ready state. (A SSD in the Failed state cannot
|
||
be initialized in ASM.)
|
||
|
||
- The Repair option in the ASM Setup program may fail to fix a
|
||
corrupted installation, depending on which files are affected.
|
||
The repair operation completes successfully, but the software
|
||
remains unfixed.
|
||
|
||
- Adaptec Storage Manager may fail to exit properly when you create
|
||
64 logical devices in the wizard. The logical devices are still
|
||
created, however.
|
||
|
||
- The "Clear logs on all controllers" action does not clear events
|
||
in the ASM Event Viewer (GUI). It clears device events, defunct
|
||
drive events, and controller events in the controllers' log files.
|
||
To clear events in the lower pane of the GUI, select Clear
|
||
configuration event viewer from the File menu.
|
||
|
||
- Stripe Size Limits for Large Logical Drives:
|
||
|
||
The stripe size limit for logical drives with more than 8 hard
|
||
drives is 512KB; for logical drives with more than 16 hard
|
||
drives it is 256KB.
|
||
|
||
- Agent Crashes when Hot-Plugging an Enclosure:
|
||
|
||
With one or more logical drives on an enclosure, removing
|
||
the enclosure cable from the controller side may crash
|
||
the ASM Agent.
|
||
|
||
--------------------------------------------------------------------
|
||
(c) 2012 PMC-Sierra, Inc. All Rights Reserved.
|
||
|
||
This software is protected under international copyright laws and
|
||
treaties. It may only be used in accordance with the terms
|
||
of its accompanying license agreement.
|
||
|
||
The information in this document is proprietary and confidential to
|
||
PMC-Sierra, Inc., and for its customers' internal use. In any event,
|
||
no part of this document may be reproduced or redistributed in any
|
||
form without the express written consent of PMC-Sierra, Inc.,
|
||
1380 Bordeaux Drive, Sunnyvale, CA 94089.
|
||
|
||
P/N DOC-01700-02-A Rev A |