IBM System Storage™ DS Storage Manager version 10.36.xx.13 for Linux on IBM pSeries Servers (Linux on POWER - LoP). Important: With the release of 07.36.08.00 and 07.36.12.00 controller firmware for the DS4000 and DS5000, customers are encountering a firmware bug that is causing the controllers to go into a continuous reboot cycle, which causes loss of access. There is no exposure to customers currently at 7.30.21.00, 7.15.07.00, 7.15.10.01, or 7.10.23.00. Customers who have downloaded the 07.36.08.00 or 07.36.12.00 levels of firmware, should not install this level of firmware. Customers who have installed this level of firmware should not make any configuration changes to the system and contact IBM Support immediately. For Customers who have downloaded the DS Storage Manager 10.36.65.07 code package, from the IBM Support web site, but have not yet installed firmware 7.36.08.00 or 7.36.12.00 , it is recommended that Storage Manager Firmware Upgrade Utility version 10.36.xx.07 or 10.36.xx.08 be uninstalled and version 10.36.xx.10 Firmware Upgrade Utility be installed and used to upgrade the controller firmware to 07.36.14.xx or higher. Note: The Linux on POWER (LoP) host attachment to the IBM 1722-all models (DS4300), IBM 1724-all models (DS4100), IBM 1742-all models (DS4500 and DS4400), 1814-all models (DS4200 and DS4700) and 1815-all models (DS4800) and 1818-all models (DS5300 and DS5100) requires the additional purchase of an IBM DS4000 Linux on POWER Host Kit Option. The IBM DS4000/ DS5000 Linux on POWER Host Kit options contain the required IBM licensing to attach an IBM pSeries server running Linux Operating Systems to the DS4100, DS4200, DS4300, DS4400, DS4500, DS4700, DS4800, DS5100 or DS5300 storage subsystems. Please contact your IBM service representatives or IBM resellers for purchasing information. IMPORTANT: This is NOT the correct DS Storage Manager version 10.36 host software package for the 64-bit (IA-64) or for the 32-bit (IA32/x32) xSeries servers. Please refer to the Storage Manager version 10.36 host software package for the 64-bit (IA-64) version of Linux or the Storage Manager version 10.36 host software package for the 32-bit (IA32/x32) version of Linux. NOTE TO SERVICE – Reference RETAIN # N/A (C) Copyright International Business Machines Corporation 1999, 2008. All rights reserved. US Government Users Restricted Rights - Use, duplication, or disclosure restricted by GSA ADP Schedule Contract with IBM Corp. Note: Before using this information and the product it supports, read the general information in section 6.0 "Trademarks and Notices” in this document. IMPORTANT: DS4000 Storage Manager version 9.19.xx.12 and later is required for DS4000 storage subsystem users in countries implementing the new Daylight Saving Time (DST) rules (which includes U.S., Canada, Bermuda). Storage Manager version 9.19.xx.12 or later release includes revised DST transition dates to comply with provision of the Energy Policy Act of 2005 extension of Daylight Saving Time (DST) by four weeks, beginning in year 2007. Please refer to section 1.3 Enhancements for more information. Refer to the IBM System Storage™ Support Web Site or CD for the IBM System Storage DS Storage Manager version 10 Installation and Host Support Guide. This guide along with the Storage Manager program Online Help provide the installation and support information. This guide also provides information on other DS4000/DS5000 related publications. Last Update: 03/27/2009 Please refer to corresponding Change History document for more information on new features and modifications. Products Supported ---------------------------------------------------------------- | New Model | Old Model | Machine Type | Model | |------------|-----------|--------------|------------------------| | DS5300 | N/A | 1818 | 53A | |------------|-----------|--------------|------------------------| | DS5100 | N/A | 1818 | 51A | |------------|-----------|--------------|------------------------| | DS4800 | N/A | 1815 | 82A, 82H, 84A, 84H, | | | | | 88A, 88H, 80A, 80H | |------------|-----------|--------------|------------------------| | DS4700 | N/A | 1814 | 70A, 70H, 72A, 72H, | | | | | 70T, 70S, 72T, 72S, | |------------|-----------|--------------|------------------------| | DS4500 | FAStT 900 | 1742 | 90X, 90U | |------------|-----------|--------------|------------------------| | DS4400 | FAStT 700 | 1742 | 1RX, 1RU | |------------|-----------|--------------|------------------------| | DS4300 | FAStT 600 | 1722 | 6LX, 6LU, 6LJ, 60X | | | | | 60U, 60J, 60K, 60L | |------------|-----------|--------------|------------------------| | DS4200 | N/A | 1814 | 7VA, 7VH | |----------- |-----------|--------------|------------------------| | DS4100 | FAStT 100 | 1724 | 100, 1SC | |----------- |-----------|--------------|------------------------| | DS3400 | N/A | 1726 | 41X, 42X, HC4 | |----------- |-----------|--------------|------------------------| | DS3300 | N/A | 1726 | 31X, 32X, 31E, 32E | |----------- |-----------|--------------|------------------------| | DS3200 | N/A | 1726 | 21X, 22X, HC2 | |----------- |-----------|--------------|------------------------| | N/A | FAStT 500 | 3552 | 1RX, 1RU | |------------|-----------|--------------|------------------------| | N/A | FAStT 200 | 3542 | 1RX, 1RU, 2RX and 2RU | ------------ --------------------------------------------------- ATTENTION: 1. The DS4000 storage subsystem controller firmware version 05.40.xx.xx and earlier along with its associated NVSRAM files should not be loaded into DS4000 storage subsystems connected to DS4000 EXP100 drive expansion enclosures. 2. For the DS4400 and the DS4100 storage subsystems - all models (Standard /dual controller and Single Controller), the controller firmware version 06.12.xx.xx and later must be used. 3. For the DS4300 storage subsystems with Single Controller option, the controller firmware version 05.34.xx.xx and its associated NVSRAM file must be used. Do not attempt to download the controller firmware version 06.xx.xx.xx into these DS4300 storage subsystem models with single controller option. 4. For the FAStT200 (M/T 3542-all models) and FAStT500 (M/T 3552-all models) storage subsystems, the controller firmware version 05.30.xx.xx and its associated NVSRAM file must be used. Do not attempt to download the controller firmware version 06.xx.xx.xx into these DS4000 storage subsystems. 5. For the DS4300 with Standard (base) or Turbo options and the DS4500 storage subsystems connected to the DS4000 EXP810 drive expansion enclosures, the latest controller firmware version 06.23.xx.xx (and higher) and its associated NVSRAM file must be used. It is also recommended that the DS4300 with Standard (base) or Turbo options and the DS4500 storage subsystems be upgraded to this controller firmware version even though they are not planned to have DS4000 EXP810 drive expansion enclosure attachment. 6. For the DS4700 and the DS4800 storage subsystems connected to the DS4000 EXP100 drive expansion enclosures, the controller firmware version 06.23.xx.xx (and higher) and its associated NVSRAM file must be used. It is also recommended that the DS4700 and the DS4800 storage subsystems be upgraded to this controller firmware version even though they are not planned to have DS4000 EXP100 drive expansion enclosure attachment. 7. For the DS3x00 storage subsystems, please refer to the readme files that are posted in the IBM DS3000 System Storage support web site for the latest information about their usage, limitations or configurations. http://www.ibm.com/servers/storage/support/ Linux kernels Supported: ======================== ---------------------------------------------------------------------------- | OS | Version | RDAC | | | | | |----------------------|--------------------------|--------------------------| | LoP-Redhat 4 | 2.6.9-67.EL | Yes(note a) | | update 6 (RHEL4-u6)| |rdac-LINUX-09.03.0B05.0030| |----------------------|--------------------------|--------------------------| | LoP-SLES 9 Service | 2.6.5-7.305-ppc64 | Yes(note a) | | Pack 4 (SLES9-SP4) | |rdac-LINUX-09.03.0B05.0030| |----------------------|--------------------------|--------------------------| | LoP-SLES 10 Service | 2.6.16.60-0.21-ppc64 | Yes(note a) | | Pack 2 (SLES10-SP2)| |rdac-LINUX-09.03.0C05.0030| |----------------------|--------------------------|--------------------------| | LoP-Redhat 5 | 2.6.18-92.EL5 | Yes (note b) | | update 2 (RHEL5-u2)| |rdac-LINUX-09.03.0C05.0030| ---------------------------------------------------------------------------- Notes: a. The redundant failover/failback capability is supported by the IBM DS Storage Manager Linux RDAC driver versions 9.03.0B05.0030 for SLES 9, RedHat 4 and SLES 10 (2.6 kernels), and 9.03.0C05.0030 for Redhat 5.x and SLES10 SP1 and greater. b. Redhat 5.0 is supported by the Storage Manager client only. ======================================================================== CONTENTS -------- 1.0 Overview 2.0 Installation and Setup Instructions 3.0 Configuration Information 4.0 Unattended Mode 5.0 Web Sites and Support Phone Number 6.0 Trademarks and Notices 7.0 Disclaimer ======================================================================= 1.0 Overview -------------- 1.1 Overview -------------- This IBM DS Storage Manager package version 10.36 for Linux on POWER is required for managing the DS5100 and the DS5300 with controller firmware version 07.36.xx.xx. In addition, it is also recommended for managing the following DS4000 storage subsystems: the DS4800-all models, the DS4700-all models, the DS4200-all models, the DS4100-all models, the DS4300-all models, the DS4400-all models and the DS4500-all models with controller firmware version 06.12.xx.xx or higher installed. NOTE: DS4000 controller firmware versions 6.12.xx.xx * Attachment support for EXP710 and EXP100 SATA disk drive expansion enclosures for DS4300 and DS4500. * No attachment support for EXP810 disk drive expansion enclosure. DS4000 controller firmware version 06.60.xx.xx * Attachment support for EXP710, EXP810, and EXP100 SATA disk drive expansion enclosures for DS4300 with dual controller, DS4500, DS4700 and DS4800. * Attachment support for EXP420 for DS4200 only. DS4000 controller firmware version 7.36.xx.xx * Supports DS4200, DS4700, and DS4800 only. DS5000 controller firmware version 7.36.xx.xx * Supports the DS5100 and DS5300 only. Important: Please refer to the "IBM System Storage DS Storage Manager version 10 Installation and Host Support Guide" located at http://www.ibm.com/support/ for all installation and support notes pertaining to Linux on pSeries. The IBM DS Storage Manager version 10 Linux support for pSeries host software packages, located on IBM's Support Web Site, contain the documentation and files necessary for installing the IBM DS Storage Manager Version 10.36 Client host software in Linux on pSeries(LoP) servers with SUSE SLES 9 Service Pack 4 (SLES9-SP4), RedHat EL 4 update 6 (RHEL4-u6), RedHat EL 5 update 2, and SUSE SLES 10 Service Pack 2 (SLES10-SP2) that are required to setup and manage the DS4000 storage subsystem that are connected to Linux on pSeries(LoP) servers. In addition, other host software packages are also available in the IBM System Storage™ Disk Storage Systems Technical Support web site. IMPORTANT: This release of IBM DS Storage Manager host software supports all DS4000/DS5000 storage subsystems with general released firmware versions from 05.30.xx.xx up to 07.36.xx.xx and the DS3200, DS3400 and DS3300 storage subsytems with controller firmware version 07.35.xx.xx or earlier, installed. For the latest DS4000 readme files, publications, and released code, refer to the http://www.ibm.com/support/ IBM System Storage™ Disk Storage Systems Technical Support web site. Refer to the IBM DS4000 Support Web Site for latest DS4000 controller firmware and NVSRAM files for IBM machine types 1722, 1724, 1742, 1814, 1815, 1818, 3542 and 3552. The following table shows the first four digits of the latest controller firmware versions that are currently available for various models of the DS4000/DS5000 storage subsystems. --------------------------------------------------- | DS4000 Storage | Controller firmware version | | Subsystem Model | | |------------------|--------------------------------| |------------------|--------------------------------| | DS5300 (1818) | 07.36.xx.xx | |------------------|--------------------------------| | DS5100 (1818) | 07.36.xx.xx | |------------------|--------------------------------| | DS4800 (1815) | 07.36.xx.xx | | | | |------------------|--------------------------------| | DS4700 (1814) | 07.36.xx.xx | |------------------|--------------------------------| | DS4500 (1742) | 06.60.xx.xx | |------------------|--------------------------------| | DS4400 (1742) | 06.12.xx.xx | |------------------|--------------------------------| | DS4300 Turbo | 06.60.xx.xx | | Option (1722) | | |------------------|--------------------------------| | DS4300 | 06.60.xx.xx | | Standard Option | | | (1722) | | |------------------|--------------------------------| | DS4300 Single | 05.34.xx.xx | | Controller Opt. | (for DS4300 SCU | | (SCU)(1722) | only) | |------------------|--------------------------------| | DS4200 (1814) | 07.36.xx.xx | |------------------|--------------------------------| | DS4100 (1724) | 06.12.xx.xx | | (standard dual | | | Single | | | controller Opt.)| | |------------------|--------------------------------| | FAStT500 (3552) | 05.30.xx.xx | |------------------|--------------------------------| | FAStT200 (3542) | 05.30.xx.xx | --------------------------------------------------- Note: Controller firmware 06.23.xx.xx or later is required for the DS4300-standard- dual controller or Turbo models, the DS4500-all models, the DS4700- all models and the DS4800-all models for attaching combinations of EXP810 and EXP100 storage expansion enclosures behind a DS4000 storage subsystem. New features introduced with controller firmware version 07.36.xx.xx or later are not available for DS4000 controllers with older firmware versions installed. The following table describes controller firmware versions required for attaching various models of DS4000 Storage Expansion Enclosures. ------------------------------------------------------------------------------- | Controller | EXP Storage Expansion Enclosures | | FW Version |----------------------------------------------------------------| | | EXP100 | EXP420 | EXP500 | EXP700 | EXP710 | EXP810 | EXP5000 | |------------|---------|--------|--------|--------|--------|--------|---------| |5.3x.xx.xx | No | No | Yes | Yes | No | No | No | |------------|---------|--------|--------|--------|--------|--------|---------| |5.40.xx.xx | No | No | Yes | Yes | No | No | No | |------------|---------|--------|--------|--------|--------|--------|---------| |5.41.xx.xx | Yes | No | No | No | No | No | No | |------------|---------|--------|--------|--------|--------|--------|---------| |5.42.xx.xx | Yes | No | No | No | No | No | No | |------------|---------|--------|--------|--------|--------|--------|---------| |6.10.0x.xx | Yes | No | Yes | Yes | No | No | No | |------------|---------|--------|--------|--------|--------|--------|---------| |6.10.1x.xx | Yes | No | Yes | Yes | Yes | No | No | |------------|---------|--------|--------|--------|--------|--------|---------| |6.12.xx.xx | Yes | No | Yes | Yes | Yes | No | No | |------------|---------|--------|--------|--------|--------|--------|---------| |6.14.xx.xx | Yes | No | No | No | Yes | No | No | |------------|---------|--------|--------|--------|--------|--------|---------| |6.15.xx.xx | Yes | No | No | No | Yes | No | No | |------------|---------|--------|--------|--------|--------|--------|---------| |6.16.2x.xx | No | No | No | No | Yes | Yes | No | |------------|---------|--------|--------|--------|--------|--------|---------| |6.16.8x.xx | No | Yes | No | No | Yes | Yes | No | |------------|---------|--------|--------|--------|--------|--------|---------| |6.16.9x.xx | No | Yes | No | No | Yes | Yes | No | |------------|---------|--------|--------|--------|--------|--------|---------| |6.19.xx.xx | Yes | No | No | No | Yes | Yes | No | |------------|---------|--------|--------|--------|--------|--------|---------| |6.23.xx.xx | Yes | Yes | No | No | Yes | Yes | No | |------------|---------|--------|--------|--------|--------|--------|---------| |6.60.xx.xx | Yes | Yes | No | Yes | Yes | Yes | No | |------------|---------|--------|--------|--------|--------|--------|---------| |7.10.xx.xx | Yes | Yes | No | No | Yes | Yes | No | |------------|---------|--------|--------|--------|--------|--------|---------| |7.15.xx.xx | Yes | Yes | No | No | Yes | Yes | No | |------------|---------|--------|--------|--------|--------|--------|---------| |7.30.xx.xx | No | No | No | No | No | Yes | Yes | |------------|---------|--------|--------|--------|--------|--------|---------| |7.36.xx.xx | Yes | Yes | No | No | Yes | Yes | Yes | ------------------------------------------------------------------------------- Note: EXP500 and EXP710/EXP100 can not be attached behind the same DS4000 storage subsystem. In addition, the EXP500 can only be intermixed with EXP700 if the fibre channel speed of the EXP700 and the DS4000 subsystem drive loop is set to 1Gbps. ======================================================================= 1.2 Limitations ---------------- IMPORTANT: The listed limitations are cumulative. However, they are listed by DS4000/DS5000 storage subsystem controller firmware and Storage Manager host software releases to indicate which controller firmware and Storage Manager host software release that they were first seen and documented. New limitations with Storage Manager Upgrade Utility package version 10.36.xx.13 release (controller firmware 07.36.xx.xx). 1. None New limitations with Storage Manager Upgrade Utility package version 10.36.xx.07b release (controller firmware 07.36.xx.xx). 1. Upgrade of DS4000 to 7.xx.xx.xx using the upgrade tool with non-English OSes will end with an error. The current firmware column still shows the previous version, pending version will show 7.xx.xx.xx, instead of 'none'. Checking the upgrade tool log you will find: [12.11.2008 08:51:16] [rey-ds4700-1] [SUPPORT_SERVICES] [DownloadAndActivate] activation failed There are two ways to check the content of this log: 1. As long as the upgrade tool is still open: Use the 'View log' button 2. If the upgrade tool was closed already: For windows all logs can be found in this folder: C:\Program Files\IBM_DS4000FirmwareUpgrade\client For other operating systems, check in the appropriate folder. Syntax of the name is similar to 20081114_1501.log. Then check the profile. Here the new firmware can be found as pending: Current configuration: Firmware version: 06.60.17.00 NVSRAM version: N1814D470R916V17.dlp Pending configuration: Staged firmware download supported: Yes Firmware version: 07.36.12.00 NVSRAM version: N1814D47R1036V12.dlp This shows that firmware was loaded to controller but not yet activated. Workaround: After checking the items listed in Details section, activate the loaded firmware using this script command: activate storageSubsystem firmware; Mark the affected system in the Enterprise Management window of the Storage Manager. In the menu, go to Tools -> Execute script. A new window will show up. Paste the command in the upper part of this window and choose Tools -> Execute only. The activation will take a while and the controller will reboot. New limitations with Storage Manager Installer (SMIA) package version 10.36.xx.07 release (controller firmware 07.36.xx.xx). 1. With SuSE Linux 10 SP1 and OES2 installed, IO operations fail when you move cluster resources. You can avoid this problem if you quiesce the system before moving cluster resources. 2. With RHEL 5.2 and the 8G QLogic HBA, the in-distro driver ignores the data underrun in the event of the Check Condition (Recovered Error) – When the HBA driver sees the command has a Check Condition status, it ignores the data underrun condition. It then passes whatever the data that is in its memory along with the Check Condition status to the failover driver. Since the failover driver treats Recovered Error as a good status and the data in the memory is incorrect, the application will report a data miscompare. The data on the disk is correct. This is reported as bugzilla 465741 - [QLogic 5.2.z bug] qla2xxx - Additional residual-count corrections during UNDERRUN handling. Limitations with Storage Manager Installer (SMIA) package version 10.30.xx.xx release (controller firmware 07.30.xx.xx). 1. Host IO errors can occur when using Emulex HBAs connected to a Cisco MDS 9124 fabric switch in a large SAN (greater than 16 servers). 2. With RHEL 5.1 and QLogic HBAs, rebooting server after adding new LUNs can fail to come up. Either run "hot_add" utility provided with the Linux MPP driver or update your QLogic driver to version 8.01.07.15-2. 3. Executing "hot_add" during IO activity can cause an IO error. This occurs when a LIP is issued and the HBA FW never returns the ready status. A patch is available from Redhat associated with bugzilla 413211. 4. Heavy IO stress with a server running RH 5.1 kernel 2.6.18-53 can cause a kernel panic. Upgrading to errata kernel 2.6.18-53.1.14 can avoid this. 5. Native Cluster is not supported with Redhat 5.1 on LoP servers. Limitations with Storage Manager Installer (SMIA) package version 10.15.xx.xx release (controller firmware 07.15.xx.xx). 1. Reconfiguration operation (DRM) is delayed under some circumstances. When a drive fails before the DRM completes, it can take up to eight times as long for the DRM to complete. DRM reconfigurations to RAID 6 have the longest impact since there are four times as many calculations and writes that have to occur compared to other RAID levels. 2. Host Software display of controller current rate is wrong below 4Gbps. Host Software Client - AMW - Logical/Physical View - Controller Properties - Host Interfaces - Current rate is displaying "Not Available" when the controller negotiated speed is 2Gbps. When the controller negotiated speed is reduced to 1GBs, the Current rate displays "2Gbps". Fixed from 10.10.xx.xx 30. When a disk array is in a degraded state, the array will report "needs attention" to both the EMW and the AMW. After taking appropriate corrective action, the AMW view of the array will report "fixing" but the EMW state remains at "needs attention". Both statuses are valid, when the fault state is resolved both views will change to "optimal". Limitations with Storage Manager Installer (SMIA) package version 10.10.xx.xx release (controller firmware 07.10.xx.xx). 1. Controller Alarm Bell icon does not appear as a flashing icon indicator on the screen to get the user's attention but the icon does change appearances. 2. Miswire of drive tray cabling with DS4700 and DS4200 can cause continuous reboot of a controller. To correct this situation, power down the subsytem, cable the drive trays correctly, and power the subsystem back up. 3. Controller button may appeared enabled and mislead the user that a controller is selected where in fact, a controller was not highlighted for the botton to appear ready. 4. Search key is not marked correctly in that page due a JavaHelp Bug with JavaHelp 2.0_01. A search for keyword "profile" ended with phase "prese 'nting p' rofile" being marked. 5. storageArrayProfile.txt should be renamed as storageSubsystemProfile.txt in Support Data. 6. Bullets Incorrectly Placed in Volume Modification Help Page 7. Unable to Escape out of Help Display. User will be required to close the window by using window close procedure (exit, etc.) 8. Bullets and Descriptions not alligned into same line in "Viewing mirror properties". 9. The Help window is not getting refreshed properly when using the AMW.Help window. Workaround is to close and reopen SANtricity. 10. CLI command failure for creation of volume(s) if capacity parameter syntax is not specified. A space will need to be used between the integer value and the units used in the capacity option of this command or, "create volume volumeGroup[4] capacity=15 GB…". 11. Customer will see high ITW counts displayed in GUI (RLS feature) and logs (files) for diagnostics (support bundle, DDC, etc.) and may be concerned that he has a problem. This will not cause a Critical MEL event. Known problem previously restricted, when a DS4700 or DS4200 controller reboots, these counters increment. 12. Single tray powercycle during IO activity causes drives in tray to become failed. Customer may see loss of drive (failed) due to timing issue of drive detection & spin-up of the drive. There are one of two conditions that result on power-up: - (Most likely) Drive will be marked as optimal/missing with the piece failed, or - (rarely) Drive will be marked as failed with the piece failed. Workaround is to unfail (revive) drive which restarts reconstruction of all pieces. 13. Event log critical Event 6402 was reported after creating 64 mirror relations. Eventually, the mirror state transitions to synchronizing and proceeds to completion on mirror creation. Workaround is to ignore the MEL logging since this occurs on creation of mirror volumes. 14. Reconfiguration operations during host IO may result in IO errors when arrays contain more than 32 LUNs. These operations include Dynamic Capacity Expansion, Defragmentation, Dynamic Volume Expansion, Dynamic RAID Migration. The workaround is to quiesce host IO activity during reconfiguration. 15. Heavy IO to a narrow volume group of SATA drives can result in host IO timeouts. A narrow volume group refers to an array built of very few drives; namely 1 drive RAID 0, 1x1 RAID 1, and 2 + 1 RAID 5. The workaround is to build arrays of SATA drives out of 4 + 1 or greater. 16. When managing the storage subsystem in-band, the upgrade utility will show the upgrade as failed. This is because of the update and reboot of the controllers when activating the new firmware. SMagent is not dynamic and will need to be restarted to reconnect to the storage subsystem. 17. Selecting and dragging text within the storage profile window causes the window to be continuously refreshed. Work around is to select and copy, do not drag the text. 18. When configuring alerts through the task assistant, the option stays open after selecting OK. The window only closes when the cancel button is selected. 19. The Performance Monitor displays error messages when the storage subsystem is experiencing exception conditions. The performance monitor has a lower execution priority within the controller firmware than responding to system IO and can experience internal timeouts under these conditions. 20. Critical MEL event (6402 - Data on mirrored pair unsynchronized) can occur under certain circumstances with synchronous RVM. The most likely scenario is when both primary and secondary are on a remote mirror and an error occurs with access to that host. Resynchronization should occur automatically, when automatic resynchronization is selected for a mirror relationship. However if any of the host sites should go down during this interval, recovery by the user is required. 21. A persistent miswire condition is erroneously reported through the recovery guru even though the subsystem is properly wired. The frequency of occurrence is low and is associated with an ESM firmware download or other reboot of the ESM. The ESM that is reporting the problem must be reseated to eliminate the false reporting. Not all miswire conditions are erroneous and must be evaluated to determine the nature of the error. 22. Drive path loss of redundancy has been reported during ESM download. This occurs when a drive port is bypassed. In some instances this is persistent until the drive is reconstructed. In other cases it can be recovered through an ESM reboot (second ESM download, ESM pull and replace). 23. Unexpected drive states have been observed during power cycle testing due to internal controller firmware contention when flushing MEL events to disk. The drives have been observed as reconstructing or replaced when they should have been reported as failed. Also volume groups have been reported degraded when all drives were assigned and optimal. An indication that this is the situation would be when drive reconstruction has not completed in the expected amount of time and does not appear to be making any progress. The work around is to reboot the controller owning the volume where the reconstruction has stalled. 24. Sometimes when an ESM is inserted a drive's fault line is asserted briefly. The fault line almost immediately returns to inactive, but the ESMs may bypass the drive. In these circumstances, the administrator will have to reconstruct the failed drive. 25. After a drive fail, a manually initiated copyback to a global hot spare may also fail. The work around is to remove the failed drive and reinsert it, then the copyback should resume and complete successfully. 26. When an erronieous miswire condition occurs (as mentioned above in 21), the recovery guru reports the miswire on one controller but not on the other. In this situation, ignore the other controller and use the information supplied by the controller reporting the problem. 27. Occasionally a controller firmware upgrade to 07.10 will unexpectedly reboot a controller an extra time. This could generate a diagnostic data capture, however the firmware upgrade is always successfull. 28. When managing previous releases of firmware (06.19 and prior), "working" gets displayed as "worki" during volume creation. 29. The Performance Monitor error window does not come to the front. You must minimize all other foreground windows to get to the error popup window. 30. When a disk array is in a degraded state, the array will report "needs attention" to both the EMW and the AMW. After taking appropriate corrective action, the AMW view of the array will report "fixing" but the EMW state remains at "needs attention". Both statuses are valid, when the fault state is resolved both views will change to "optimal". 31. Adding a storage subsystem through the SMClient using in-band management causes the server to hang with RH 5.0. You must use out-of-band management with RH 5.0. 32. SCSI fencing does not work when set up through the Redhat Conga web interface. Use system-config-cluster utility instead. 33. For SLES 9 SP3, the SCSI driver does not create a device handle for the UTM LUN. Do not define a UTM LUN or use in-band management with this release. 34. RHEL 5 SMIA package installation can fail with library exceptions because of a dependency on deprecated libraries. The following libraries must be installed before installing the SMIA: libXp-1.0.0.8 compat-libstdc++-33-3.2.3 (non zSeries) compat-libstdc++-296-2.95.3 (zSeries) or install all packages on the RHEL 5 host prior to installing SMIA. 35. SLES10 SP1 on pSeries model 655 with EOD LSI Logic / Symbios Logic 53c1010 66MHz Ultra3 SCSI adapter fails to reboot properly after installing SMIA version 10.10. The work around is to remove SMagent and not allow in-band management. 36. Hot swap HBA fails on SLES9 SP3 with LP11000 adapter. Port 1 of the HBA fails after completing the hot swap. 37. Configuring separate Email alerts when two Enterprise management windows are open on the same host will cause the alerts to disappear if one of the Enterprise windows is shut down and then restarted. It is recommended that if multiple Enterprise management windows needs to be open, that they are open on separate Hosts which will indeed allow the configuration of alerts to be saved if one of the enterprise management windows is shut down and restarted. Legacy restrictions that are still applicable: 1. Reflected Fibre Channel (FC) OPN frames occurs when intermixing the EXP810, EXP710 and EXP100s behind DS4700 or DS4800 storage subsystems. This behavior causes excessive drive side timeout, drive side link down and drive side link up events be posted in the DS4000 storage subsystem Event log (MEL.) It might also cause drives to be by-passed or failed by the controller. NEW DRIVE SIDE FC CABLING REQUIREMENT MUST BE ADHERED TO WHEN HAVING EXP100 CONNECTED TO THE DS4700 OR DS4800 STORAGE SUBSYSTEMS. Please refer to the latest version of the Installation, User's and Maintenance Guide for these storage subsystems that are posted in the IBM DS4000 Support web site for more information. http://www.ibm.com/servers/storage/support/disk/ 2. Can not increase the capacity of RAID arrays. RAID arrays with certain combinations of selected segment size and number of drives that made up the arrays will exceed the available working space in controller dacstore, causing a reconfiguration request (like expanding the capacity of the array) to be denied. These combinations are generally the largest segment size (512KB) with the number of drives in the array is 15 drives or more. There is no work-around. C324144 105008 3. Interoperability problem between the Tachyon DX2 chip in the DS4500 and the DS4300 storage subsystem controllers and the Emulex SOC 422 chip in the EXP810 expansion enclosure ESMs causing up to 5 Fibre Channel loop type errors to be posted in the DS4000 storage subsystem Major Event Log during a 24 hour period. There is a small window in the SOC 422 chip that multiple devices can be opened at one time. This ultimately leads to Fibre Channel loop errors of Fibre Channel link up/down, Drive returned CHECK CONDITION, and Timeout on drive side of controller. IBM recommends the use of the Read-Link-Status function to monitor drive loop for any problems in the drive loop/channel. There is no work-around. 4. The single digit of the Enclosure IDs for all enclosures (including the DS4000 storage subsystem with internal drive slots) in a given redundant drive loop/channel pair must be unique. For example, with four enclosures attached to the DS4300, the correct enclosure ID settings should be x1, x2, x3, and x4 (where x can be any digit that can be set). Examples of incorrect settings would be 11, 21, 31, and 41 or 12, 22, 32, and 62. These examples are incorrect because the x1 digits are the same in all enclosure IDs (either 1 or 2). If you do not set the single digit of the enclosure IDs to be unique among enclosures in a redundant drive loop/channel pair, then drive loop/channel errors might be randomly posted in the DS4000 subsystem Major Event Log (MEL), especially in the cases where the DS4300 storage substems are connected to EXP810s and EXP100s. In additon, enclosure IDs with same single digits in a redundant drive loop/channel pair will cause the DS4000 subsystem controller to assign a soft AL_PA address to devices in the redundant drive loop/channel pair. The problem with soft AL_PA addressing is that AL_PA address assignment can change between LIPs. This possibility increases the difficulty of troubleshooting drive loop problems because it is difficult to ascertain whether the same device with a different address or a different device might be causing a problem. 5. In DS4000 storage subsystem configurations with controller firmware 6.15.2x.xx and higher installed, the performance of write intense workloads such as sequential Tape restores to DS4000 Logical drives with large I/O request sizes (e.g. 256kB) is degraded if the DS4000 logical drives are created with small segment sizes such as 8KB or 16KB. The work around is to create the DS4000 logical drives with segment size of 64KB or higher. 6. Do not pull or insert drives during the drive firmware download. In addition, ALL I/Os must also be stopped during the drive firmware download. Otherwise, drives may be shown as missing, unavailable or failed. 7. Do not perform other storage management tasks, such as creating or deleting logical drives, reconstructing arrays, and so on, while downloading the DS4000 storage subsystem controller firmware and DS4000 EXP ESM firmware. It is recommended that you close all storage management sessions (other than the session that you use to upgrade the firmware) to the DS4000 storage subsystem that you plan to update. 8. The "modprobe -r" command is not supported on RHEL3 when the Linux RDAC driver is installed. Use "rmmod" to remove and "modprobe" to recover the device driver stack. On pSeries servers with more then 4 processors, using the modprobe command may result in a hung server. 9. The Storage Manager Linux RDAC driver does not have the functionality to detect when a LUN is removed from the server. This includes FLASH Copy LUN's. Error messages are reported to the /var/log/messages file until missing LUN is recognized. The work around at this time is to reboot the pSeries server. 10. Universal Access Luns should not be assigned to LUN ID 0, especially with RDAC's installed. Placing an Universal Access Lun at LUN ID 0 may lead to loss of LUN recognition of next sequence LUN IDs or partial list of Virtual Luns reported with the RDAC driver. 11. SAN Boot is not supported on IBM JS20/JS21 server blades with the initial release SLES9 kernel. The SAN Boot is only supported on JS20 server blades with SLES9 w/SP1 kernel and above. (6-3-2005) 12. Uninstalling the storage manager host software package might leave the JAVA JRE directory behind. This problem might occur if you uninstall a copy of the SANtricity Storage Manager software that has been upgraded from a previous release. The newer installation package could not delete the JRE directory created with an older installation package. The workaround is to manually delete any directories left behind after uninstalling the software. 13. Linux 2.6 kernels (RHEL4 and SLES9) with Emulex HBAs require device driver version 8.0.16.26 or higher. 14. Utilizing standard Emulex installation script (lpfc-install) on SAN booted systems with RDAC installed will remove the currently used lpfc module, causing loss of access to root filesystem. Installation script uses usage count of lpfc module to determine if it is appropriate to remove the module. Installing RDAC driver will cause usage count of lpfc module to go to zero. Work-around Perform the following steps a. Boot system without RDAC driver installed. b. Update Emulex driver and reboot. c. Install RDAC driver and reboot 15. Because the MPP driver installation sets physical HBA module options during installation in /opt/mpp/modprobe.conf.mppappend. The following procedure is recommended when updating the HBA driver. a. Get the latest hba drivers supported by MPP. b. install the drivers c. execute "mppUpdate" to create a new mpp initrd image. d. reboot the system to boot with the new initrd image. See the Readme.txt file distributed with the MPP driver for more information. 16. On System p servers running RHEL4u4, an issue exists in the kernel where a panic can occur on the host system when the system is under heavy IO load. There is currently no work-a-round for this issue. RedHat is aware of the issue and has a proposed patch that will resolve the problem. Customers should contact RedHat and refer to RedHat defect number 116302 (bz32990_rh116302_cmvc360422_cmvc360556_cr117379_cr117456). 17. On System p servers running RHEL4u4 an issue exists where host system IO processes will stop making progress, pending IO completion from the kernel. This can be seen by looking at IO processes runtime where the process is outstanding for a longer time then expected. To work-a-round this issue, the SMdevices command being run when the condition is present, has shown to allow the hung IO processes to proceed. RedHat is aware of the issue and has a proposed patch that will resolve the problem. Customers should contact RedHat and refer to RedHat defect number 116303 (bz32992_rh116303_cmvc363570_cr118958). 18. When a Linux on Power system is booted on the SAN, where the boot/root volume is associated to a DS4000 Logical Volume, an issue was found were, if LILO is used to update the /etc/yaboot.conf boot loader information, while the boot/root LUN is under the control of the mpp driver, the result will be an invalid /etc/yaboot.conf file and lilo throwing exceptions during execution. To avoid this issue, do not use, LILO, YaST, or any other application to update the /etc/yaboot.conf bootloader configuration file after initial configuration. A work-a-round, is to manually edit the /etc/yaboot.conf, and perform the updates as needed to modify the bootloader configuration file in place; normally the /etc/lilo.conf file would be edited, and then lilo would be run to perform the update. Refer to the Linux documentation, manual, and info pages for a details on how to edit the file. If a kernel needs to be installed or upgrade, or if a driver is installed or upgraded, it is possible that LILO will still be run by the install scripts. In this case, if the mpp driver is still loaded, then this issue will be encountered. It is recommended that a backup copy of the /etc/yaboot.conf be kept in case of this occurring. An alternative option is to reboot the system into single path mode, without the mpp driver being loaded, before any updates are done to the kernel and/or drivers, to help to avoid this issue. If this is done, then normal applications can be used to update the boot loader configuration files. 19. On System p servers running SLES 10, an issue exists in the kernel where a panic can occur on the host system when the storage subsystem does a failover. The following kernel PTFs are available from Novell under the SR #10325172063. http://you.novell.com/update/ppc/update/SUSE-SLES/10/PTF/014fcd0641ec128c68ddfc2ce730090f/20070315 http://you.novell.com/update/i386/update/SUSE-SLES/10/PTF/014fcd0641ec128c68ddfc2ce730090f/20070315-1 http://you.novell.com/update/x86_64/update/SUSE-SLES/10/PTF/014fcd0641ec128c68ddfc2ce730090f/20070315-2 ======================================================================= 1.3 Enhancements ----------------- DS Storage Manager version 10.36 host software for Linux operating systems includes support for controller firmware 07.36.xx.xx: - Maintenance Release for DS5300, DS5100, DS4800, DS4700, and DS4200. See change history file for detail list of field fixes. - Increased the number of supported Host Types from 16 to 32. - Storage Manager v10.36 Event Monitor installation default changed from disabled to enabled. - Additional MEL Health Checks have been included in the Storage Manager Firmware Upgrade Utility; > MEL entries have wrapped during the last 24 hours > More than 10 DDE Critical MEL events reported in the last 7 days > More than 2 Page 2C Critical MEL events reported in the last 7 days > More than 1 Unscheduled Controller Reboot Critical MEL evens (0x5005, 0x5040) reported in the last 7 days > More than 2 Degraded Drive Channel Critical MEL events reported in the last 7 days > More than 8 Host side Channel Error Critical MEL events reported in the last 7 days > Non-redundant Drive Side cabling on the drive side > More than 4 critical MEL entries in the last 7 days > Add check for event showing MEL was cleared and record this in the Utility Log. Please refer to the New Features section of the IBM System Storage DS Storage Manager Version 10 Installation and Host Support Guide for additional information about the IBM DS Storage Manager version 10 enhancements. ======================================================================= 1.4 Prerequisites ------------------ Note: The IBM Storage Manager host software version 10.36 for Linux on Power OS new features and changes are described in the corresponding Change History document. Please refer to this document for more information on new features and modifications. (also see Dependencies section) Code levels at time of release are as follows ------------------------------------------------------------------ IMPORTANT: There are three separate IBM DS Storage Manager host software version 10.36 packages for Linux operating system environments. Host software packages exist for each of the following Linux operating systems 1. 32-bit X86 and 64-bit x86_64 or X64 (AMD64 or EM64T). 2. Intel Itanium II 64-bit. 3. Linux on Power (LoP). Please use the correct host software package for your Linux operating system environment. Starting with the DS4000 storage manager version 9.16, all of the host software packages (SMruntime, SMclient, SMesm, SMutil and SMagent) are included in a single DS4000 Storage Manager host software installer wizard. During the execution of the wizard, the user will have a choice to install all or only certain software packages depending on the need for a given server. Note: The Linux RDAC package is not part of this installer wizard. It must be installed separately, if needed. For Linux servers without the graphics adapter, individual host software installation packages are provided in the DS Storage Manager Version 10.36 host software package. These individual host software installation packages can also be downloaded at the IBM System Storage™ Disk Storage Systems Technical Support web site. If you wish to use the wizard but do not have a graphics adapter, you can execute the installer via sh -i console to run in console mode. The version of the host software installer wizard for this release is SMIA-LINUXPPC-10.36.A5.13.bin. This installer wizard will install the following version of the host- software packages SMruntime: 10.30.A5.01-1 SMclient: 10.36.G5.09-1 SMesm: 10.36.G5.02-1 SMutil: 10.00.A5.06-1 SMagent: 10.00.A5.05-1 Note: The SMagent package can be installed when the Linux RDAC is configured as multipath (Failover/Failback) driver. It should not be installed if the FC HBA Linux failover device driver is used as multipath driver. When using Linux RDAC driver as the multipath failover/failback driver, the non-failover version of the Linux FC HBA device driver must be installed instead of the failover version. The SMFWUpgInstaller-LINUXPPC-10.36.A5.10.bin will install the following: SMfwupgrade 10.36.G5.05-1 SMfirmware 10.36.G5.08-1 The minimum DS4000 controller firmware version that the Linux RDAC driver supported was 05.4x.xx.xx. Refer to the README that is part of the Linux RDAC device driver package for installation instructions. The latest version of these drivers are as follows: - IBM DS4000 storage subsystem Linux RDAC: 09.00.A5.22 (for 2.4 kernel only) rdac-LINUX-09.00.A5.22-source.tar.gz 09.03.0B05.0030 (for 2.6 kernel only) rdac-LINUX-09.03.0B05.0030-source.tar.gz 09.03.0C05.0030 (for RH 5.0 and SLES10SP1 and greater) rdac-LINUX-09.03.0C05.0030-source.tar.gz ======================================================================= 1.5 Dependencies ----------------- 1. The IBM System Storage DS4000 Controller Firmware Upgrade Tool is required to upgrade any system from 6.xx controller firmware to the 7.36.xx.xx controller firmware. 2. Always check the README files (especially the Dependencies section) that are packaged together with the firmware files for any required minimum firmware level requirements and the firmware download sequence for the DS4000 drive expansion enclosure ESM, the DS4000 storage subsystem controller and the hard drive firmware. 3. DS4000 storage subsystem controller firmware version 07.15.xx.xx requires that the firmware version of an attached EXP100 environmental service module (ESM) be at version 9566. It is not possible to upgrade the ESM firmware from version 953X to version 956X or later directly. The 953X ESM must be upgraded to version 9554 first before it can be upgraded to version 956X. 4. Standard installation order for Storage Manager 10.36.xx.xx and controller firmware 07.36.xx.xx: 1. SMruntime - always first 2. SMesm - required by client 3. SMclient 4. SMagent 5. SMutil 6. Controller firmware and NVSRAM 7. ESM firmware 8. Drive firmware IBM DS Storage Manager version 10.36 host software requires the DS4000/DS5000 storage subsystem controller firmware be at version 05.30.XX.XX or higher. The only IBM DS4000 Storage Manager supported on DS4000 storage subsystems with controller firmware version 04.xx.xx.xx up to 05.2x.xx.xx is version 09.60.xx.xx. ======================================================================= 1.6 Level Recommendation and Prerequisites for the update ----------------------------------------------------------- IMPORTANT: Linux 2.6 kernels (RHEL4 and SLES9) with Emulex FC HBAs require Emulex device driver version 8.0.16.27 or higher. At this time, the supported drivers are as follows: SLES 9 SP4 QLogic 8.01.07.15 Emulex 8.0.16.34 SLES 10 SP2 QLogic 8.00-k6 Emulex 8.2.0.22 RHEL 4 U6 QLogic 8.01.07-d4 Emulex 8.0.16.34 RHEL 5.2 QLogic 8.02.00-k5 Emulex 8.2.0.22 The Emulex device drivers for Linux on Power can be downloaded at the following web page: http://www.emulex.com/ibm/support/index.jsp The QLogic device drivers for Linux on Power can be downloaded at the following web page: http://driverdownloads.qlogic.com/QLogicDriverDownloads_UI/IBM.aspx?companyid=6 One can also use a non-pSeries Server or a pSeries server running AIX OS as the management station. Please refer to the appropriate operating system readme file for code levels. These readme files can be found in the other operating system (WIN32, WS03_X86_32bit, Linux or AIX) directories on the DS Storage Manager version 10 support for Linux on POWER Host Kit CD-ROM. IMPORTANT: Always refer to the http://www.ibm.com/support/ IBM System Storage™ Disk Storage Systems Technical Support web site for the latest released code levels. It is recommended that you use host type LINUX for Linux on Power hosts connected to your storage subsystem. The failover driver recommendation is MPP/RDAC which requires that AVT be disabled. As with all NVSRAM changes from the default, anytime you download new NVSRAM you must re-enable the changes that you have made. In this case, you will have to re-run the AVT disable scripts. ========================================================================== 2.0 Installation and Setup Instructions ---------------------------------------------------- Note: The web-download Storage Manager version 10.36 host software package to support Linux on Power (LoP) servers must be first unpacked (tar - xvz) into a user-defined directory. Then, go to this directory, locate the LoP directory to access the Storage Manager version 10 host software installation file(s). IMPORTANT: There are three separate IBM DS Storage Manager host software version 10.36 packages for Linux operating system environments. Host software packages exist for each of the following Linux operating systems 1. 32-bit X86 and 64-bit x86_64 or X64 (AMD64 or EM64T). 2. Intel Itanium II 64-bit. 3. Linux on Power (LoP). Please use the correct host software package for your Linux operating system environment. 2.1 Step-by-step instructions for this code update are as follows ----------------------------------------------------------------------------------------- 1. If needed, install and update the driver for the IBM DS4000 Host Adapter. a. Install the hardware by using the instructions that come with the adapter. b. Install the IBM DS4000 Host Adapter driver by using the instructions provided in the readme file located in the HostAdapter directory on the installation CD or downloaded with the device driver source package from the IBM support web site. c. Install the IBM DS4000 Linux RDAC driver by using the instructions provided in the readme file located in the LinuxRDAC directory on the installation CD or downloaded with the device driver source package from the IBM support website. 2. If there is a previous version 7.x, 8.x or 9.xx of the IBM DS4000 Storage Manager host software (ie. SMRuntime, SMClient, RDAC, SMUtil and SMAgent packages) installed in the server, you have to uninstall it first before installing the new version of the storage manager host software. 3. Install the new Storage Manager host software version from the CD-ROM directory or from the host software package that you downloaded from the IBM Support Web site. Refer to the IBM System Storage DS Storage Manager version 10 Installation and Host Support Guide for detailed installation instructions. CAUTION: Always check the README files (especially the Dependencies section) that are packaged together with the firmware files for any required minimum firmware level requirements and the firmware download sequence for the DS4000/DS5000 drive expansion enclosure ESM, the DS4000/DS5000 storage server controller and the hard drive firmware. Refer to the http://www.ibm.com/support/ IBM System Storage™ Disk Storage Systems Technical Support web site for the the latest DS Storage Manager host software, the DS4000/DS5000 controllers firmware, the drive expansion enclosure ESM firmware and the hard disk drive code. 2.2 Helpful Hints ----------------------- 1. The DS4500 and DS4300 storage subsystem have updated recommended drive-side cabling instructions. The DS4500 instructions are documented in the IBM System Storage DS4500 Installation, Users, and Maintenance Guide (GC27-2051-00 or IBM P/N 42D3302). The DS4300 instructions are documented in the IBM System Storage DS4300 Installation, Users, and Maintenance Guide (GC26-7722-02 or IBM P/N 42D3300). Please follow the cabling instructions in these publications to cable the new DS4500 and DS4300 setup. If you have an existing DS4500 setup with four drive side minuhub installed that was cabled according to the previously recommended cabling instructions, please schedule down time as soon as possible to make changes to the drive side FC cabling. Refer to the IBM System Storage DS4500 and DS4300 Installation, Users. and Maintenance Guide for more information. 2. Running script files for specific configurations. Apply the appropriate scripts to your subsystem based on the instructions you have read in the publications or any instructions in the operating system readme file. A description of each script is shown below. - EnableAVT_Linux.script: The script will enable automatic logical drive transfer (ADT/AVT) for the Linux heterogenous host region. Do not use this script unless it is specifically mentioned in the applicable instructions. (This script can be used for other host type if modifications are made in the script, replacing the Linux host type with the appropriate host type that needs to have AVT/ADT enabled) - DisableAVT_Linux.script: The script will disable the automatic logical drive transfer (ADT/AVT) for the Linux heterogenous host region. Do not use this script unless it is specifically mentioned in the applicable instructions. (This script can be used for other host type if modifications are made in the script, replacing the Linux host type with the appropriate host type that needs to have AVT/ADT disabled) - EnableCntlReset.script: This script is to set the HostNVSRAMByte offset 0x1B of the LnxclVMWare host type NVSRAM region to 1 for enabling the propagation of the bus/target/LUN reset that one controller received to the other controller in a dual controller DS4K subsystem. - DisableCntlReset.script: This script is to set the HostNVSRAMByte offset 0x1B of the LnxclVMWare host type NVSRAM region to 0 for disabling the propagation of the bus/target/LUN reset that one controller received to the other controller in a dual controller DS4K subsystem. The default propagation of the bus/target/LUN reset setting for the LnxclVMWare host type is to enable it. This script should be used only when the following conditions are met. . The DS4000 storage subsystem did not have any LUNs mapped to any Linux hosts that are part of Linux cluster configuration. . The DS4000 storage subsystem did not have any LUNs mapped to any VMWare hosts. . The Linux hosts use Linux RDAC as multipath driver. . The host type of defined host ports/group is LnxclVMWare(LNXCL). Inappropriate use of this script might cause loss of access to the mapped LUNs from the VMware hosts or Linux hosts that are part of a Linux Cluster. 3. The ideal configuration for SATA drives is one drive in each EXP per array, one logical drive per array and one OS disk partition per logical drive. This configuration minimizes the random head movements that increase stress on the SATA drives. As the number of drive locations to which the heads have to move increases, application performance and drive reliability may be impacted. If more logical drives are configured, but not all of them used simultaneously, some of the randomness can be avoided. SATA drives are best used for long sequential reads and writes. 4. IBM recommends at least one hot spare per EXP100 drive expansion enclosure. 5. If you are unable to see the maximum number of drives during Automatic Configuration, you should be able to use Manual Configuration to select individual drives and select the maximum number of drives allowed. 6. The DS4000 controller host ports or the Fibre Channel HBA ports can not be connected to a Cisco FC switch ports with "trunking" enabled. You might encounter failover and failback problems if you do not change the Cisco FC switch port to "non-trunking" using the following procedure: a. Launch the Cicso FC switch Device Manager GUI. b. Select one or more ports by a single click. c. Right click the port(s) and select Configure, a new window pops up d. Select the "Trunk Config" tab from this window, a new window opens e. In this window under Admin, select the "non-trunk" radio button, it is set to auto by default. f. Refresh the entire fabric. 7. Serial connections to the DS4000 storage controller must be set to a baud rate of either 38400 or 57600. Do not make any connections to the DS4000 storage subsystem serial ports unless it is instructed by IBM Support. Incorrect use of the serial port might result in lost of configuration, and possibly, data. 8. Starting with the DS4000 Storage Manager (SM) host software version 9.12 or later, the Storage Manager client script window looks for the files with the file type of ".script" as the possible script command files. In the previous versions of the DS4000 Storage Manager host software, the script window looks for the file type ".scr" instead. (i.e. enableAVT.script for SM 9.12 or later vs. enableAVT.scr for pre-SM 9.12) 9. Do not delete the Access LUN or Access Volume if you want to manage the DS4000 storage subsystem in-band (host-agent managed). The Access LUN is required by the SMClient to communicate with the storage controllers when using the in-band management method. 10. Fabric topology zoning requirement with AIX fcp_array (RDAC) and Solaris RDAC only. To avoid possible problem at the host level, it is best practice that all Fibre Channel (FC) Switches must be zoned such that a single FC host bus adapter can only access one controller per storage array. In addition, this zoning requirement also ensures the maximum number of host connections can be seen and log into the controller FC host port. This is because if a FC HBA port is seen by both controller A and B host ports, it will be counted as two host connections to the storage subsystem - one for controller A port and one for controller B port. Note: The DS4000 storage subsystems DS4500, DS4400 and FAStT500 (IBM machine type 1742 and 3552) have two ports per controller - one per minihub slot. The DS4000 storage subsystems DS4300 (IBM machine type 1722) and DS4100 (IBM machine type 1724) have two ports per controller. The DS4000 storage Server FAStT200 (IBM machine type 3542) has only one port per controller. The DS4700 storage subsystem (IBM machine type 1814) has up to four ports per controller. The DS4800 storage subsystem (IBM machine type 1815) has four ports per controller. 11. All enclosures (including DS4000 storage subsystem with internal drive slots) on any given drive loop/channel should have complete unique ID's, especially the single digit (x1) portion of the ID, assigned to them. For example, in a maximum configured DS4500 storage subsystem, enclosures on one redundant drive loop should be assigned with id's 10-17 and enclosures on the second drive loop should be assigned with id's 20-27. Enclosure id's with the same single digit such as 11, 21 and 31 should not be used on the same drive loop/channel. In addition, for enclosures with mechanical enclosure ID switch like DS4300 storage subsystems, EXP100 or EXP710 storage expansion enclosures, do not use enclosure ID value of 0. The reason is with the physical design and movement of the mechanical enclosure ID switch, it is possible to leave the switch in a “dead zone” between ID numbers, which return an incorrect enclosure ID to the storage management software. The most commonly returned enclosure ID is 0 (zero). In addition to causing the subsystem management software to report incorrect enclosure ID, this behavior also result in enclosure ID conflict error with the storage expansion enclosure or DS4000 storage subsystem intentionally set the ID to 0. The DS4200 and DS4700 storage subsystems and the EXP420 and EXP810 storage expansion enclosures do not have mechanical ID switches. Thus, they are not susceptible to this problem. In addition, these storage subsystems and storage expansion enclosures automatically set the Enclosure IDs. IBM recommendation is not make any changes to these settings unless the automatic enclosure ID settings resulting in non- unique single digit settings for enclosures (including the storage subsystems with internal drive slots) in a given drive loop/channel. ======================================================================= 3.0 Configuration Information ------------------------------------------ 3.1 Configuration Settings ------------------------------------- 1. When using the Linux RDAC as the multipath driver, the host type can be set to either LNXCLVMWARE(Linux Cluster) or LINUX host type. If the chosen host type is Linux, one must run the DisableAVT_Linux.script to disable the Auto logical drive transfer feature. If the DisableAVT_Linux.script is not run, the mapped LUNs might not failover/failback properly when using Linux RDAC as the multipath driver. If Linux RDAC is not used as multipath driver, the only host type that can be chosen is Linux. In this case, DO NOT run the DisableAVT_Linux.script to turn off the Auto logical drive transfer feature. Instead, run the EnableAVT_Linux.script to ensure that the Auto logical drive transfer feature is enable. 2. By default, the IBM DS Storage Manager Client does not automatically map logical drives when the IBM DS4000 storage partitioning premium feature is enabled. This means that the logical drives after being created are not automatically presented to the host systems. a. For a new installation, after creating new arrays and logical drives, if your host OS type is any of the supported version of Linux OS's, create a partition with the host type of Linux and map the logical drives to this partition. b. If you are upgrading the NVSRAM with Storage Partitions, you may have to change the default host type to match the host system OS like LINUX or LNXCL(LNXCLVMWARE). After upgrading the NVSRAM, the default host type is reset to Windows 2000/Server 2003 non-clustered for storage subsystem with controller firmware version 06.15.xx.xx or later. For DS4000 storage server with controller firmware version 06.12.xx.xx or earlier, it is reset to Windows non-clustered (SP5 or higher), instead. 3. If a host in a cluster server configuration lost a physical path to a storage subsystem controller, the logical drives that are mapped to the cluster group will periodically failover and then failback between cluster nodes until the failed path is restored. This behavior is the result of the automatic logical drive failback feature of the RDAC multipath driver. The cluster node with a failed path to a DS4000/DS5000 controller will issue a failover command of all logical drives that were mapped to the cluster group to the controller that it can access. After a programmed interval, the nodes that did not have failed path will issue failback command for the logical drives because they can access the logical drives both controllers, resulting in the cluster node with the failed path not be able to access certain logical drives. This cluster node will then issue a failover command for all logical drives, repeating the logical drives failover-failback cycle. The workaround is to disable this automatic failback feature. For Linux cluster (SteelEye) environments, perform the following steps: a. Open the /etc/mpp.conf file, and change the DisableLunRebalance parameter to 3. Save the changes. b. From a shell prompt, type the following command, and press Enter: mppUpdate c. Reboot the computer for the changes to take effect. d. Repeat this procedure on any system in the cluster configuration that has the Linux RDAC driver installed. ======================================================================= 3.2 Unsupported Configurations -------------------------------------------- The configurations that are currently not being supported with IBM DS4000 Storage Manager Version 10.36 are listed below: 1. The DS4100 (machine type 1724-all models) storage subsystem does not support the attachment of the DS4000 EXP710, EXP700 and EXP500 (FC) drive expansion enclosure. 2. The DS4800 storage subsystem (machine type 1815-all models) does not support the attachment of the FAStT EXP500 and DS4000 EXP700 drive expansion enclosures. 3. The DS4200 (machine type 1814 - models 7VA/H) does not support the attachment of the DS4000 EXP100 (SATA), EXP710 (FC) and EXP810 (SATA and FC) drive expansion enclosures. In addition, it does not support Fibre Channel disk drive options. 4. The IBM DS4000 EXP420 Expansion Enclosure is not supported attached to any other IBM DS4000 Storage Controllers except the DS4200. 5. The DS4100 with Single Controller option does not support the attachment of the DS4000 storage expansion enclosures. 6. The DS5100 and DS5300 storage subsystems do not support the attachment of the DS4000 EXP100, EXP700, EXP710 drive expansion enclosures. 7. The DS5000 EXP5000 drive expansion enclosure is supported attached to the DS5100 and DS5300 only. 8. The DS4700 and DS4800 storage subsystems do not support the attachment of the DS4000 EXP700 drive expansion enclosures. The EXP700 enclosure must be upgraded into DS4000 EXP710 enclosure using the DS4000 EXP700 Models 1RU/1RX Switched-ESM Option Upgrade Kit before it can be attached to the DS4700 and DS4800 storage subsystems. 9. The DS4300 storage subsystem with Single Controller option does not support the controller firmware version 06.xx.xx.xx. The correct firmware version for these DS4300 storage subsystem models is 05.34.xx.xx. 10. Fibre Channel loop environments with the IBM Fibre Channel Hub, machine type 3523 and 3534, in conjunction with the IBM Fibre Channel Switch, machine types 2109-S16, 2109-F16 or 2109-S8. In this configuration, the hub is connected between the switch and the IBM Fibre Channel RAID Controllers. 11. The IBM Fibre Channel hub, machine type 3523, connected to IBM machine type 1722, 1724, 1742, 1814, 1815, 3542 and 3552. 12. A configuration in which a server with only one FC host bus adapter connects directly to any DS4000 storage subsystem with dual controllers is not supported. The supported configuration is the one in which the server with only one FC host bus adapter connects to both controller ports of any DS4000 storage subsystem with dual controllers via Fibre Channel (FC) switch (SAN-attached configuration.) ======================================================================= 4.0 Unattended Mode --------------------- N/A ======================================================================= 5.0 WEB Sites and Support Phone Number ---------------------------------------- 5.1 IBM System Storage™ Disk Storage Systems Technical Support web site: http://www.ibm.com/support/ 5.2 IBM System Storage™ Marketing Web Site: http://www.ibm.com/servers/storage/disk 5.3 If you have any questions about this update, or problem applying the update go to the following HelpCenter World Telephone Numbers URL: http://www.ibm.com/planetwide 5.4 The IBM Linux for pSeries web site: http://www.ibm.com/eserver/pSeries/linux IMPORTANT: You should download the latest version of the DS Storage Manager host software, the DS4000/DS5000 storage subsystem controller firmware, the DS4000/DS5000 drive expansion enclosure ESM firmware and the drive firmware at the time of the initial installation and when product updates become available. For more information about how to register for support notifications, see the following IBM Support Web page: www.ibm.com/systems/support/storage/subscribe/moreinfo.html You can also check the Stay Informed section of the IBM Disk Support Web site, at the following address: www.ibm.com/support/ ======================================================================= 6.0 Trademarks and Notices ---------------------------- 6.1 The following terms are trademarks of the IBM Corporation in the United States or other countries or both: IBM FAStT System Storage™ the e-business logo xSeries pSeries HelpCenter Microsoft, Windows, and Windows NT are trademarks of Microsoft Corporation in the United States, other countries, or both. Java and all Java-based trademarks and logos are trademarks or registered trademarks of Sun Microsystems, Inc. in the United states, other countries, or both. UNIX is a registered trademark of The Open Group in the United States and other countries. Linux is a registered trademark of Linus Torvalds. Other company, product, and service names may be trademarks or service marks of others. ======================================================================= 7.0 Disclaimer ---------------- 7.1 THIS DOCUMENT IS PROVIDED "AS IS" WITHOUT WARRANTY OF ANY KIND. IBM DISCLAIMS ALL WARRANTIES, WHETHER EXPRESS OR IMPLIED, INCLUDING WITHOUT LIMITATION, THE IMPLIED WARRANTIES OF FITNESS FOR A PARTICULAR PURPOSE AND MERCHANTABILITY WITH RESPECT TO THE INFORMATION IN THIS DOCUMENT. BY FURNISHING THIS DOCUMENT, IBM GRANTS NO LICENSES TO ANY PATENTS OR COPYRIGHTS. 7.2 Note to U.S. Government Users -- Documentation related to restricted rights -- Use, duplication or disclosure is subject to restrictions set forth in GSA ADP Schedule Contract with IBM Corporation.