Professional Documents
Culture Documents
Clar PSP 084ndu
Clar PSP 084ndu
REVISION HISTORY:
3/10/03 Initial release to support the GA of Release 11 software of
4/04/03 Reformatted for clarity regarding the need to add Utility partition and
recover image during the update.
4/15/03 Added step to remove Snap Cache LUNs from the Snap Cache LUN pool
during the NDU process of updating core to Release 11
5/7/2003 Removed references to CDs for Recovery images.
06/16/2003 Update procedure and array based revisions of software.
06/18/2003 Add note on Primus case for using a LAN cross-over cable for NDU
Introduction
This document describes the procedure for updating CX-series arrays based
software using the Non-Disruptive Upgrade (NDU) capability of the array.
Please read all the rules and comments about updating a CX-series array
based software. During this procedure, each Storage Processor will reboot
independently. Therefore, it is imperative that all servers have access to both
Storage Processors in the array and have properly configured/working fail-over
software or they will lose access to the array. It is important to ensure that I/O
is at a very low level when this update is performed. Times of heavy I/O should
be avoided. The revision level of software running on the array before the
update can impact the process. Reference Table 1 to understand current and
previous array based software revision levels. The update procedure differs if
the revision of software, which the array is running, is earlier than Release 11
(R11). If updating from a release prior to Release 11 (definition below), you
must load the “Utility Partition. The Utility partition allows for recovery of a
corrupted boot image for SPA or SPB.
CLAR-PSP-084 Page 1 of 23
Updating software on a CX Series Array
Definitions
Flare Software available prior to Release 11
o CX200 – Core SW 02.03.x.20.5.xxx 6.2 Navisphere Array
components (release 10 Cx200 only)
o CX400 – Core SW 02.02.x.40.5.xxx 6.2 Navisphere Array
components (release 8)
o CX600 – Core SW 02.02.x.60.5.xxx 6.2 Navisphere Array
components (release 8)
o CX600 – Core SW 02.01.x.60.5.xxx 6.1 Navisphere Array
components (release
NDU Software Update Types (visible in the NDU dialogue box during the
process.)
9 Disruptive – Updating an array with one functioning Storage
Processor. This is standard for a CX200LC but is only possible
if in engineering mode for all other array types via the GUI.
Navisphere CLI will prompt you for a confirmation.
9 Non-Disruptive – Updating an array that has 2 functioning
Storage Processors. Each Storage Processor updates
individually, always leaving one working on the array. If all
Servers have valid fail-over software, this typical update will not
disrupt server access to data on the array.
9 Rebootless – This type of Software update will reboot neither
storage processor. A Server will not be affected at all during
this procedure.
9 Utility Partition – Software loaded to previously unused but
reserved space on the first 4 drives of a CX-Series array that
provides menus/utilities to perform recovery and conversion of
an array.
9 Repository – A previously reserved but un-used space on
drive 0_4 (5th drive) in the boot chassis that holds Recovery
images that are NDU’d to the array or added via the FTP
method.
Q When is it loaded
A All arrays shipped since March 10th 2003 have the Utility partition pre-
loaded to each of the first 4 drives. In addition, there is a recovery image
preloaded at the factory. This Image is loaded to private space on drive 4
called the “repository”
Q Why do some files have a .NDU extension and some have .MIF
A The Utility partition is available only as a .NDU file and must be loaded via
an NDU operation. The Recovery image is available as a .NDU file and a
.MIF. The .MIF is the version that is FTP’d directly from your laptop to the
SP. See clar-psp-078.
9 RULE - Do Not Start an NDU operation if there are any LUN rebuilds, disk
equalizations, or Raid Group/LUN expansions in progress.
9 RULE - View the Event Log to confirm that the array has not been
experiencing recent problems which should be investigated and understood
prior to this NDU.
9 RULE – Load a Utility partition via NDU If updating from a release prior to
Release 11. This is done AFTER the array is running Release 11 minimum.
9 RULE All connected servers must have working and supported “path-
failover” software in order to perform this procedure. If they do not, stop and
prevent all I/O from that server to the array
9 RULE - Confirm that there is at least one working path from each
connected server to each Storage Processor of the array
9 RULE - You cannot NDU a new revision of code to the array if the previous
revision requires a commit. Perform a commit on the “active” version if the
Array software Tab shows that a “commit” is required.
9 RULE - All array based software within a “release” must be updated at the
same time unless stated otherwise. See the Software revision table in this
document.
9 RULE - CX600, CX400 and CX200 arrays should have less than 100 I/Os
per second to each of the first 4 drives in the array (drives 0_0 through
0_3). This low I/O activity to the boot drives ensures that the NDU will have
no time-out problems and will finish successfully. See appendix A for
instructions on measuring them I/O rate to the first 4 drives. This procedure
may not be necessary if there is little or no I/O running.
9 RULE - Following an NDU where Navisphere Management Server or the
Manager UI is revised, you MUST close your browser and re-open it in
order to use the new content that may be in the new revisions of
Navisphere Management server or UI.
9 RULE - Navisphere agents and CLI on Servers must be no more than one
revision behind the SP based Navisphere agent.
9 RULE – Navisphere CLI (being used to manage) must be equal to or one
revision greater than any Navisphere Host or SP Agent it will manage
9 RULE - Host based AdmSnap software must be at the same revision as
SnapView running in the array
9 RULE - If updating FROM a version prior to Release 11, add the Utility
Partition and the Recovery Image via a second NDU AFTER the array has
been update to Release 11 or higher.
9 RULE - Stop all SAN Copy sessions before starting an NDU
9 RULE - Stop all Snap Sessions prior to NDU
9 RULE - Move Snap Cache LUNs out of the Snap Cache pool prior to NDU.
9 RULE – CLARalert software must be at the same release level as the SP
Agent
9 RULE – The Navisphere Agent used by the CLARalert monitoring (dial-
home) workstation must be at the same level as the SP Agent
9 RULE – Access Logix must be already loaded on the array before
SnapView or MirrorView, can loaded for the first time.
9 RULE - The Software Installation Wizard is for the CX200 and CX200LC
only.
Notes &
Warnings about
NDU
General comments
NDU updated to SnapView 2.10.06 from a LAN crossover cable to one SP.
Notes for specific
attached host
environments
AIX Servers
• Stop all I/O from any connected AIX server during the NDU process if
it is running the native HBA driver. See primus emc67186
Solaris Servers
• Ensure that the no-device-delay setting is configured to '0' (zero) in
lpfc.conf. See PRIMUS emc66969
• NDU Delay - The Software Installation window of the Navisphere
Manager GUI has a check-box for NDU delay. When checked, the
default for time will be 360 seconds. When there is a Server zoned to
the array that has DMP fail-over software, use the default of 360
seconds.
MirrorView
• Do not “deactivate” a mirror prior to an NDU. In fact, deactivating the
mirror will cause all server access to the primary LUN to be lost until the
Mirror is reactivated.
• Multiple Arrays involved in a MirrorView Configuration should not be
targeted during the same software NDU operation you should perform
the NDU on one Array at a time.
• Provided the software on the arrays is within 2 revisions, no special
preparation of the mirrors is required prior to an NDU. The array
software ensures that mirrors continue to operate correctly during an
NDU, without requiring a full re-synchronization of any secondary
image. System fractures are a normal part of the NDU operation, but
these result in partial re-synchronization only (which will automatically
start if the mirror recovery policy is set to “Automatic”)
Clones
• Do not start an NDU operation while there are more than 4 reverse
synchronization operations active on the array.
San Copy
• All San Copy sessions should be stopped prior to NDU.
NDU Procedure 1. Repair any Faults on the array - Confirm there are no faults on the array
before you begin
2. Confirm valid working failover software and access to both SPs for
each Server with access to the array.
3. Ensure you have new versions of all array based software currently
running.
4. Ensure there are no LUN Rebuilds, Equalizations or RG/LUN
Expansions in progress. If there are, wait for them to complete.
5. Download the appropriate Recovery image and Utility partition (if
required) before starting the update.
6. Put all files to be updated, into the same directory on your Service laptop
or server you will be using to run Navisphere Manager. See the software
compatibility table in this document. As a general rule, all array-based
software must be updated at the same time via the same NDU.
7. Before SW update Measure I/O to the first 4 drives of the array
(CX200, CX400 and CX600) and ask the customer to reduce the I/O if
possible. There should be no more than 100 I/Os per sec to the first 4
drives.
8. Stop all snap sessions
9. Move all Snap Cache LUNs out of the Cache LUN Pool
10. Connect to the array using Navisphere Manager.
11. Enter Engineering mode (note: the Software Installation Wizard can be
used only for the CX200 and CX200LC)
12. Perform the Software update
• Right-click the array icon
• Select the array properties and the array Software tab of the
properties dialogue.
• Choose the update option and browse to the files to be updated
• Load the selected files.
13. Monitor status via the status window which automatically opens. When
the status “operation has completed successfully” appears, the update is
complete.
14. Confirm that the new SW is listed as ACTIVE via the Software tab of the
array properties dialogue box.
15. COMMIT the new Base (Flare or Access Logix) code.
16. Refresh the VIEW of the Browser. This step is necessary to begin using
the new Navisphere UI loaded to the array.
17. If Power path is running and LUNs are not listed under their default
owner, perform a powermt restore.
19. Return Snap Cache LUNs to the Cache pool If previously removed and
restart snap sessions as needed.
“Loading it via FTP at this time will be disruptive to the customer since you
must reboot one of the SPs and use it to add the recovery image to the
array. If this activity is not appropriate for this time it can be added to the
array at the time of need, without further interruption to the customer.”
NOTES:
o See Section in this Document titled “Utility Partition FAQs
o See Appendix B in this document for instructions for using the
Utility Partition and loading the recovery Image via the FTP
method.
o Refer to CLARiiON procedure clar-psp-078 for further
instruction on adding and using a recovery image.
21. Update CLARalert If installed and test call home
22. Confirm that there are no array faults (the SPS units may still be charging
based upon the SP reboots, this is normal).
23. Confirm that Cache Settings are set properly. If new layered applications
are being added, Cache setting may not be able to be set as they were
previously.
• If updating from one release to the next, update all the below “array-based” software via the same
software update
• Arrays shipped from the factory with Release 11 code on them had the utility partition and recovery
Image pre-loaded. Reloading the Utility Partition and Recovery Image is NOT required if updating from
Release 11 to Release 11 patch.
• There was not a release 9
• Latest Versions are noted below in RED
Access
Logix
CX400 Flare Not released 02.02.0.40.5.xxx Not revised 02.04.0.40.5.002 02.04.0.40.5.005 02.04.0.40.5.007
Access
Logix
CX200 Flare Not released Not released CX200 introduced 02.04.0.20.5.002 02.04.0.20.5.005 02.04.0.20.5.007
02.03.0.20.5.xxx
CX200 Not released Not released CX200 introduced 02.04.1.20.5.002 02.04.1.20.5.005 02.04.1.20.5.007
02.03.1.20.5.xxx
Access
Logix
SnapView 1.5 2.0 Not released 2.1 2.10.04 2.10.06
Not released Not released 01.00.02 01.01.01 Not revised Not revised
HA Enabler
Navisphere
Components 6.1.0 6.2.0 6..2.0 6.4.0 & 6.41 UIs Not revised Not revised
(The
Management
Svr/Agent went to
6.2.1)
NOTES
1. Host Navisphere Agents follow the array based agent revision (one rev back on the host agent is
acceptable)
2. Host based AdmSnap revisions must follow the revision of SnapView
3. Navisphere UIs revised to 6.4.1 after Rel 11 GA and before 11+ GA
4. All Release 11 components are compatible, but the latest versions should always be run together
5. HA enabler is CX200 Only
6. Thirty-drive option is CX200 only (std for EMC customers)
7. Navi Base components for CX200 only
8. It is NOT required to update the Utility partition or the Recovery Image when going from Release 11 to
Release 11 .007.
To ensure the success of NDU operations on CX200, CX400 and CX600 systems, the I/O level on first
four disk drives (FRUs 0, 1, 2, 3) should be measured prior to executing an NDU operation.
• If the I/O level to any of these disk drives is higher than 100 IOs per second, the NDU should be
postponed or the IO load to the LUs based on these CRUs should be reduced until the I/O load
falls below these thresholds.
• The NDU may Time-out and fail if any of the first 4 drives are performing greater than 100 IOs per
second.
For each of these disk drive selectors, right click on the selector, then select
"Properties". Select the "Statistics" tab in the new window. Add together the "Read
Throughput (IO/Second)" and "Write Throughput (IO/Second)" statistics to produce
the current number of I/O requests per second for the indicated CRU (disk drive). If
the values are displayed as "unavailable", go back to step one above to enable
gathering performance statistics.
The I/O requests per second load for the first four disk drives should each average
under 100 I/O requests per second on a CX200, CX400 or CX600.
When complete, ensure that the “check boxes” in the Array Properties dialogue box
which “enable statistics logging” for each SP are returned to the way they were prior
to this procedure.
1. Execute the Navi CLI command from a server with LAN or PPP access to the array.
navicli -h <SP-x address> getdisk 0_0_0
This gives Read and Writes since boot of the array. Note the results
4. Subtract the original cumulative read and write values returned in step one from the
values returned in step three and then add the two differences together to produce
the total number of I/Os performance for the wait period.
5. Divide the total number of I/Os by the wait period (in seconds) to produce the number
of IOs per second.
The five steps above must be executed twice - once for SPA and once for SP-B for a
given disk and the number of IOs per second added together to produce the total number
of IOs per second for the disk drive.
The five steps above must also be executed on both SPs for all of the first four disk drives
(0_0_0, 0_0_1, 0_0_2, and 0_0_3).
The I/O requests per second load for the first four disk drives should each average under 100 I/O
requests per second on a CX200, CX400 or CX600.
If the I/O load exceeds the recommended limits listed above, it must be reduced for the LUs (disk
Logical Units) based on these disk drives. The Navisphere Manager windows and Navisphere
CLI commands used above to gather the statistics data display the LUNs based on these disk
drives.
When Complete, return the Statistics logging capability to off
The following procedure can be done to load the Recovery image to an array if the Utility partition is already loaded but the
Recovery image is not loaded or it fails to load via NDU with an error message similar to “attempt to store image to
persistent storage failed” .
1. Connect a serial cable from a Windows server/desktop/laptop. The cable connects from a Comm port on the computer to
the Serial Port on SPA. Use the characteristics below.
The Comm Port on the computer should be set to;
9600 baud
8-bit
No parity
1 stop-bit,
No flow control.
2. Open a hyper-terminal session on the Windows server/desktop/laptop using the port characteristics above.
3. While monitoring the hyper-terminal session, reboot SPA by entering Engineering mode and re-booting SPA. Note: SPB
Can be in and running I/O at this time.
4. The Storage Processor will begin its boot sequence running BIOS and POST Tests. While monitoring the power-up
sequence of the Storage Processor be prepared to interrupt this process by hitting the Esc Key on your keyboard at
anytime during the printout of this string of Alpha characters as noted in the figure below.
DB_key
5. Following the printout of the text “Call your service Representative” type the characters DB_key followed by an “enter”
on the keyboard. It is case sensitive
6. This will display the Diagnostic Menu a screen similar to the following figure
8. Note : The menu options and numbers will likely vary from product to product or revision to
Revision but “Image Sub-Menu” will always be available
9. The Image Submenu will be displayed as shown below. Select “Utility Partition Boot”. This means the SP you are
attached to will boot from the Utility partition of that alternate SP’s boot drives. The reason for this is to provide an
operating system that will allow the system to perform the tasks associated with system recovery.
10. You should then see the screen below. Press enter to continue to the Clariion Utility Toolkit Main Menu. If this screen
does not appear, the Utility Partition for this SP is not bootable. At this point, we have two options: 1) Attempt to boot
the peer SP to its Utility Partition, or 2) Go to Scenario 2 at the beginning of this section, and install a Utility Partition
Image and a Recovery Image via NDU.
11. Select Enable LAN Service Port from the following menu
12. Enter the characteristics of the LAN Port of the SP which you are physically connected to via the serial cable.
14. After you answer Y from the previous Window, Open a browser or FTP Client software package. Type in the LAN
address of the SP you are connected to via the serial cable. . In the example below it is ftp://10.14.1.174. A Login
window will open as displayed below. The Username is clariion, the password is clariion!
15. Open a Windows Explorer window, Drag and drop or move/copy the Recovery .mif file from the CD noted earlier in
this document or previously downloaded to your service laptop, to the large FTP window just opened.
16. When the File is loaded to the FTP Window (actually loaded to a temporary area known as the RAM Disk) go back to
the Hyperterminal Window which is still open. Select Image Repository Sub-menu
17. From the above screen, select Copy files from RAM disk to Repository. This copied the recovery file from the SP
memory to the repository on Drive 0_4.
19. The above screen confirms that the file has been copied from RAM Disk to the repository. Go back to the main menu
and select RESET SP-A
21. Remove the Serial Cable from SPA and Ensure that SPA boots as normal. Confirm via Navisphere that the SP is running
the correct SW. Perform a commit of the Base SW if the SW Status Tab in Navisphere states that a commit is required.
END