Professional Documents
Culture Documents
IBM FlashSystem HyperSwap Cookbook v1.8
IBM FlashSystem HyperSwap Cookbook v1.8
Configuration, Architecture
Guidelines and Operational Best
Practices
Authors:
Anil Narigapalli
Gavin O’Reilly
IBM FlashSystems HyperSwap Configuration, Architecture Guidelines and Operational Best Practices
Contents
IBM FlashSystem HyperSwap Overview and Configuration 04
Overview 05
IP Quorum 06
HS Configuration requirements and guidelines 07
IP Quorum Deployment 10
HyperSwap Topology Configuration Steps 13
HyperSwap Tuning Parameters 16
Contents
Test Case 3: IP Quorum Failure 36 GTS Standard Building Capacity Guidance 57
Test Case 4: Inter Site Communication Failure 37 IBM FS9200 and FS7200 Capacity Sizing Guidance 58
Test Case 5: Host/App Node Failure in Site 1 38 IBM FS9200 and FS7200 Capacity Sizing Guidance - Explained 59
Test Case 6: Host/App Node Failure in Site 2 39 HyperSwap Change Volumes Capacity Considerations 60
Test Case 7: Complete Site 1 Failure 40
Test Case 8: Complete Site 2 Failure 41
Test Environment Configuration Sample Guidance 42
Overview
• HyperSwap (HS) is the high availability (HA) solution for IBM • In MPC IaaS for Storage environments, certain accounts warrant the
FlashSystems and previous Storwize systems such as FS5100, need for NAT’ing the node canister IPs to communicate with the IP
FS7200, FS9100, FS9200, V7000 etc. that provides continuous data quorum servers deployed in the ISPW cloud which is treated as a
availability in case of hardware failure, power failure, connectivity third failure domain in the HS topology. Generic topology of the
failure, or site failure, disasters. IBM FlashSystem system (e.g. FS9200) HS configuration with IP
quorum is depicted below:
• HS uses intra-cluster synchronous remote copy (Metro Mirror)
capabilities. HS essentially makes a host’s volumes accessible
across two IBM FlashSystem/Storwize system I/O groups in a
clustered system by making the primary and secondary volumes of
the Metro Mirror relationship, running under the covers, look like
one volume to the host.
• This section details the HS considerations, requirements, logical
configuration of the HS functionality with IP Quorum in IBM
FlashSystems and Storwize storage systems.
• This section also details the process for the systems which use
Network Address Translation (NAT) IPs for the IBM FlashSystem,
Storwize system node canister service IPs.
• Host configurations are not included in this section. Follow the host
configuration guidelines for HS deployment when using HS
topology.
5 © 2018 IBM Corporation February 16, 2024 IBM Services
IBM FlashSystem HyperSwap Overview and Configuration
IP Quorum
• A quorum device is used to break a tie when a SAN fault occurs, when exactly half of the nodes that were previously a member of the system are present.
A quorum device is also used to store a backup copy of important system configuration data. An IP quorum application is used in Ethernet networks to
resolve failure scenarios where half the nodes or enclosures on the system become unavailable.
• HyperSwap implementations require FC storage or an IP Quorum on a third site to cope with tie-breaker situations if the inter-site link fails, and when
connectivity between sites 1 and 2 is lost. In MPC IaaS for Storage deployments, wherever feasible ISPW location will be used a third site to host the IP
quorum application. This site is treated as an independent and a third failure domain. IP quorum is a java-based application and is specific to a cluster.
• Key considerations while using IP Quorum are as follows:
Never deploy IP quorum in primary and/or secondary sites.
Deploy a minimum of 2 IP quorum instances in third failure domain i.e. site 3.
Do not deploy the IP quorum application on a host that depends on storage that is presented by the IBM HS FlashSystem (e.g. FS9200).
When certain aspects of the system configuration change, an IP quorum application must be reconfigured and redeployed to hosts. These aspects
include:
– Adding or removing a node from the system.
– When service IP addresses are changed on a node.
– Changing the system certificate or experiencing an Ethernet connectivity issue.
– If an IP application is offline, the IP quorum application must be reconfigured because the system configuration changed.
– When a firmware upgrade happens.
• For MPC IaaS for Storage deployments, when the quorum application is hosted in ISPW location, it is preferred to configure the IP quorum application
without a quorum disk for metadata due to bandwidth constraints.
• Note: Follow the physical planning and SAN connectivity and zoning guidelines as outlined in the IBM knowledge center. Details on physical
planning and SAN connectivity, zoning for HS topology is not covered in this section.
IP Quorum Deployment
• This section details the process to deploy an IP quorum in third site • The deployment steps for an IBM FlashSystem/Storwize system
i.e. ISPW site which is treated as a third failure domain independent (e.g. FS9200, applicable to other IBM FlashSystems/Storwize
of primary and secondary sites. This section covers the details for systems such as FS5100, FS7200, FS9200, V7000 etc.) are as
the deployment for the environments where the NAT’ed IPs are follows:
used for IBM FlashSystems/Storwize system (e.g. FS9200) node
canister service IPs and for the environments with no NAT in place. 1. Login to the FS9200 cluster GUI.
2. Traverse to System -> Settings -> Network -> Service IPs.
• Certain MPC IaaS for Storage environment deployments use
3. (Applicable only if NAT’ed IPs are used) Modify the service
NAT’ed IPs for FS9200 node canister service IPs. As the quorum
IPs of all the node canisters to reflect the corresponding
application generated on an IBM FlashSystems/Storwize (e.g.
NAT’ed IP of the source service IP.
FS9200) cluster embeds the service IPs into the application, it is
essential to ensure that the NAT’ed IPs are fed into the quorum
application while it is generated on the IBM FlashSystem/Storwize
(e.g. FS9200) clustered system. Else the communication from
quorum server to the IBM FlashSystems/Storwize system (e.g.
FS9200) service IPs will not happen. It is not allowed to change the
node service IPs in the quorum application manually.
IP Quorum Deployment
4. Generate and download the IP Quorum application. d. Download the IP quorum application by clicking the option
a. Traverse to Settings -> System -> IP Quorum. “Download IPv4 Application”
b. Click on “Quorum Setting” and change the “mode” and “site”
as shown below:
c. The above step will set the quorummode to “preferred” and e. Select the option “Download the IP quorum application
the preferred site as “site_mo1” which means that in case of a without recovery metadata”.
tiebreaker scenario, the preferred site will be allowed to
resume the I/O operations. “SITE_MO1” is mentioned for
illustration purpose only. Change it according to the
environment.
IP Quorum Deployment
f. The above option is selected due to the network bandwidth 8. Change the permissions for the IP quorum application. The
limitations from ISPW site to the production sites. As the application file will generally be named as “ip_quorum.jar”
quorum drives hosted on the primary and secondary FS9200
9. Execute the IP quorum jar file to start the IP quorum application.
systems host the recovery metadata. A quorum disk designated
a. nohup java -jar ip_quorum.jar &
by a managed drive in the quorum configurations contain a
b. replace ip_quorum.jar with the correct IP quorum application
reserved area that is used exclusively for system management.
file name.
In this case, the recovery metadata will be stored on these
drives. In case of system recovery scenarios, this can be used. 10. Ensure the log and lock files (.lck) are generated in the directory
g. As soon as the quorum application is created, the download is from which the jar file is executed.
initiated to the local workstation. Save a copy of the IP
11. Verify the java processes are running on the IP quorum server.
Quorum application to a specific folder on the local
a. ps -ef | grep -I java
workstation.
12. Login to the CLI of FS9200 cluster.
5. (Applicable only if the NAT’ed IPs are used for node canister
service IPs) Modify the service IPs of all the FS9200 node canisters 13. Verify the quorum status using the command “lsquorum”
back to the original IPs by replacing the corresponding NAT’ed IPs
with the original service IPs. 14. The output must list the “IP quorum” application and mark it as
“active”
6. Copy the IP Quorum application downloaded to the local
workstation to the IP quorum server.
7. Login to the IP quorum server and change to root.
Design Considerations
• HyperSwap configurations must span three independent failure domains. Two • Requirement to use GTS LCM Code guidance and ensure the FS9200s are at
failure domains contain storage systems e.g. FS9200s. Third failure domain the target level 8.3.x or higher code level
i.e. third site contains IP quorum instances. The IP quorum should never be in
• Standalone non-HyperSwap configured FS9150/FS9200/Storwize has a max
the same failure domain as the storage systems.
usable capacity of 875 TB as per GTS sizing guidelines.
• The distance between the two failure domains containing Storage should not
• HyperSwap configured FS9200 / Storwize maximum Cluster wide (2 I/O
exceed 100 KM and a round trip latency of 3ms on the fibre channel SAN
groups) capacity is 1.75 PB.
extension.
•
When migrating off large capacity environments (i.e.: SVC w/+2PB)
A minimum of one FS9200 control enclosure is required in each of the two
you can deploy multiple FS9200 HyperSwap pairs.
failure domains containing storage.
•
The maximum number of HyperSwap enabled volumes per FS9200
Deploy identical storage systems in both the sites i.e. same model, same
HyperSwap pair is 1250.
hardware specs and same capacity.
•
For a two I/O group HyperSwap system, the maximum allowed is 875
FlashCopy can be used to take point-in-time copies of HyperSwap Volumes to
TB of host data.
use for testing, cloning and backup. However, FlashCopy mapping copy
direction cannot be reversed for restoring to a volume that is in an HyperSwap Make sure to consider the 1250 limitation when defining which
relationship. Applications / Hosts will use HyperSwap volumes. You can not split an
• Application and or Hosts within a Consistency Group between different
The use of VMware vSphere Virtual Volumes (VVols) on a system that is
HyperSwap pairs.
configured for HyperSwap is not currently supported.
•
Usage of all 1250 HyperSwap relationships is dependent on no other
Host multipath driver must be configured with the ALUA multipath policy.
FlashCopy (FC) mappings use cases. If you are using FlashCopy (FC)
mappings for other use cases, the 1250 max will be reduced by a Qty of
1 for every Qty 4 of FlashCopy mapping consumed.
Design Considerations
• Bitmap space considerations: • Rolling Disaster Scenario - Important Note(s):
The sum of all bitmap memory allocation for all functions (which Access loss between Failure domains 1 and 2 (which contain the FS9200 I/O
include Remote Copy (Metro Mirror, Global Mirror, and active- groups) because of a rolling disaster. One site is down, and the other is still
active relationships), RAID, Volume Mirroring) except working. Later, the working site also goes down because of the rolling
FlashCopy must not exceed 552 MiB per IO group. disaster.
For the 2 I/O group IBM FlashSystems/Storwize systems The system continues to operate the failure domain with the IBM
completely dedicated for HS requirements, the overall storage FlashSystems/Storwize system I/O Group, which wins the quorum race. The
volume capacity must be considered carefully. cluster continues with operation, while the I/O Group node canister in the
1.75 PB virtual capacity for a 2 I/O group IBM other failure domain stops. Later, the “operational” IBM
FlashSystems/Storwize system is ideal. This implies 875 TB per FlashSystems/Storwize I/O Group is down too because of the rolling disaster.
I/O group. All IBM FlashSystems/Storwize system I/O Groups are down.
For 875 TB of HyperSwap volumes, each of the two I/O groups In this scenario, the frozen surviving site can be restarted by manually
need: triggering the “overridequorum” command which performs the manual
438 MB of remote copy bitmap space. quorum override. However, note that “overridequorum” will modify the
Around 70 MB of RAID bitmap space. revived part of the cluster (the one that lost quorum in a disaster scenario) so
20 MB volume mirroring bitmap space which can that all offline parts of the cluster will not be able to join it again. Therefore,
accommodate 40 TB of mirrored volumes. all HyperSwap relations will have to be reconfigured and rebuilt after the
This ideal volume capacity will leave around 24 MB buffer disaster recovery operations. Hence it is strongly recommended to not use
for any operations that require bitmap space. the "overridequorum" command without product support guidance and
recommendations.
• Redundant fabrics with each fabric divided into public and private dedicated SANs. • Zoning Considerations:
Enable the NPIV feature on the cluster, and zone all hosts to
• Dedicated inter-site links (ISL) for the public and private SANs.
the virtual WWPNs on each node port.
• Implement each redundant fabric on a separate provider rather than having both fabrics Use the physical WWPNs on the node ports for inter-node
on both providers. communication. Do not zone any hosts or controllers to these
• Size the inter-site links for the private SAN appropriately, as all of the writes for ports.
Use the physical WWPNs on the node ports that are used for
HyperSwap volumes traverses these links.
inter-cluster (replication) communication. Do not zone any
• All links between sites on the public and private SANs should be trunked, rather than
hosts or controllers to these ports.
using separate ISLs. Use the physical WWPNs on the node ports for controller
• For Cisco SAN, ensure to have a separate/dedicated ISL or port-channel and allow (storage) communication.
only private VSAN traffic. The private fabrics should only have a single zone each that
includes all of the cluster node ports attached to those fabrics.
• For Brocade SAN, do not allow private SAN to traverse over shared XISLs. Deploy
separate ISLs in the virtual fabrics used for Private and Public SANs.
References:
• Private fabrics must be completely private – including implementing separate ISLs
http://www.redbooks.ibm.com/redpapers/pdfs/redp5597.pdf
from public fabrics. Do not use Private SAN for any other workloads.
https://www.ibm.com/support/pages/node/6248725
occurrence of Client Impacting Events (CIEs). resource on the remote system, each node will have a specific amount of credit for every
other node/canister it can communicate with.
• SV 8.2.x and 8.3.x technically allows coexistence of
This means each node can send a certain amount of data/information to another node
HyperSwap and Remote Mirror i.e. Metro/Global Mirror
based on the credit it has available to communicate with that node. Once the credit has
between two clusters. In an enterprise storage environment,
been used new messages will queue.
there can be a need for HS and MM/GM coexistence for
Obviously, this credit also uses resource on the local node/canister. With HS systems, in
certain scenarios. Guidance in this section help the teams
order to get sufficient performance, the distribution of the credits is such each node has
“when to” and “when not to” use both the functionalities.
'a lot' of credit for communicating with nodes in the same cluster.
IP quorum traffic between storage sites and quorum site routes via the Primary
IPsec VPN to Site-1 and traverses inter-site IP network links to Site-2 under
normal operations. The Standby route only becomes active if the Primary
route is unavailable.
Primary IPsec VPN Standby IPsec VPN
Primary IPsec VPN: 1
5
1. The traffic from IP Quorum-A server hosted in Site-3 to FS9200 in
Site-1 as depicted in flows 1, 2. 3
7
2. The traffic from IP Quorum-A server hosted in Site-3 to FS9200 in
Site-2 will happen as depicted in flows 1, 3, 4 using Inter DC IP 2 6
connectivity.
4
8 Inter DC IP Connectivity
Standby IPsec VPN:
3. The traffic from IP Quorum-A server hosted in Site-3 to FS9200 in
Site-1 will happen as depicted in flows 5, 7, 8 using Inter DC IP FC Connectivity
connectivity.
IBM FS9200 - A IBM FS9200 - B
4. The traffic from IP Quorum-A server hosted in Site-3 to FS9200 in
Site-2 as depicted in flows 5, 6 using Standby IPsec VPN.
Client Site-1 Client Site-2
24 © 2018 IBM Corporation February 16, 2024 IBM Services
IBM FlashSystem HyperSwap IP Quorum Design Considerations and Best Practices
No Impact to HyperSwap Cluster
Failure of Primary IPsec VPN Availability
Cloud Site-3
Primary IPsec VPN Failure:
• When the Primary IPsec VPN fails, the connectivity between IP
Quorum-A server and FS9200s in Site-1 and Site-2 will be lost. IP Quorum A
e
tie break device.
c
Primary IPsec VPN Standby IPsec VPN
Ra
1
5
um
• Whichever half talks to the IP quorum first, and locks it, wins,
or
Qu
now has 3 votes and continues as the running site.
3
• As GTS Best Practice is for Preferred Site configured IP Quorum 7
there is a 3 seconds head start for that site. 6
2
• The non-preferred site contacts the IP quorum, sees that it is 4
8 Inter DC IP Connectivity
locked and halts any I/O through its nodes.
• In this scenario, the FS9200 systems will continue to function
only from the preferred site. Down
FC Connectivity Offline
IBM FS9200 - A IBM FS9200 - B
• If site B experiences a failure while IP quorum is unavailable, the "worst case" May
is nodes at site A may lease expire and stop processing I/O.
Down FC Connectivity Lease
• If access to the IP quorum device is restored then the nodes at the surviving Expire
site can come online again, but there will have been an I/O outage. A manual IBM FS9200 - A IBM FS9200 - B
override would only be required if the IP quorum couldn't be restored
successfully.
Client Site-1 Client Site-2
Recommendations
Cloud Site-3
• Redundant IP quorum applications
Deploy two (2) IP quorum application instances on two different VMs.
IP Quorum A IP Quorum B
Ensure physical hardware redundancy for the VMs.
The system will automatically activate the standby instance if the first
instance fails.
• Redundant network links to quorum site Primary IPsec VPN Standby IPsec VPN
1 5
Have two IPsec VPN connections - one to primary site and one to
secondary site.
3
Have inter-site IP connectivity between primary and secondary sites 7
for cross site FS9200 to IPQW communication.
2 6
Enable auto switching of the traffic from primary VPN to secondary
4
VPN and vice-versa in the instance of either of the IPsec VPN failures. 8 Inter DC IP Connectivity
• The testing is aimed at confirming the HyperSwap functionality works in specific failure scenarios. As a best practice a testing period should
be allocated in the project implementation plan. Depending if the solution is being deployed in a new greenfield or existing brownfield
environment, will determine what level of testing can be achieved. Ensure to follow the guidance closely when approaching testing is a brown
field deployment as described in the following slides as the testing method to simulate the test differs.
• Phase-1 Testing is performed by the Storage support team to prove the HyperSwap Storage infrastructure implementation has been installed
and configured as per best practice and is working as designed. Scope of components tests to include Storage unit, SAN Fabric and Quorum
testing only.
• Phase-2 Testing is performed with the Storage support team in partnership with all the Operating System and application support teams
configured to use the HyperSwap feature. This testing phase is to prove end-to-end functionality. Scope of components tests to include
Storage unit, SAN Fabric and Quorum testing only, Host OS and Application failover compatibility.
Phase-1: Pre-Go-Live Test Cases – Compulsorily to be Phase-2: Operating system and Application End to End
performed before placing in production:
HyperSwap Test Cases:
• Test Case 1: IBM FS9200 I/O group failure in site 1.
• Test Case 5: Host/App Node Failure in site 1.
• Test Case 2: IBM FS9200 I/O group failure in site 2.
• Test Case 6: Host/App Node Failure in site 2.
• Test Case 3: IP Quorum Failure.
• Test Case 7: Complete Site 1 Failure.
• Test Case 4: Inter site communication failure.
• Test Case 8: Complete Site 2 Failure.
Caution: In a Brown Field deployment no testing that could
causes outage into a pre-existing production environment can
be performed. Test cases 4,7,8 are disruptive testing scenarios
and will only be performed in a Green Field deployment
Linux
• In HA and other standard deployment scenarios, it is essential to tune the • dev_loss_tmo set to 120 seconds.
scsi related timeouts to ensure the device path failures/issues are handled This is fibre channel specific and deals with the loss of connection to remote fibre
effectively. channel ports. The timeout countdown most often begins when the fibre channel fabric
informs the HBA that a port is no longer connected to the fabric and so is no longer
• All RHEL6, RHEL7, and SLES12 systems require that you set the
reachable. (It can also begin for other events like an HBA reset or reset of the HBA's link,
“scsi_mod.inq_timeout” parameter to 70 seconds. Otherwise, RHEL6,
which will at least temporarily cause a loss of connections to ports through the HBA.)
RHEL7, and SLES12 hosts cannot regain previously failed paths such as
The lost port has until dev_loss_tmo is reached to be rediscovered or any LUNs behind
in a system update or where a node is manually rebooted.
To resolve this issue, add scsi_mod.inq_timeout=70 to the kernel
the lost port will be considered gone and any waiting commands failed. It is number of
seconds to wait before marking a link as "bad". Once a link is marked bad, IO running on
boot command line through grub configuration. By adding the
its corresponding path (along with any new IO on that path) will be failed.
scsi_mod.inq_timeout=70 parameter, the change in the parameter is
persistent from a server reboot. Linux hosts can also regain system
node paths when lost.
• Set the udev rules for SCSI command timeout.
The Linux SCSI layer sets a timer on each command. When this
timer expires, the SCSI layer will quiesce the host bus adapter
(HBA) and wait for all outstanding commands to either time out or
complete.
Afterwards, the SCSI layer will activate the driver's error handler.
Set the SCSI command timeout to 120s. This is the recommended
setting for all versions of Linux.
References:
• https://www.ibm.com/support/knowledgecenter/STSLR9_8.3.1/com.ibm.fs9200_831.doc/svc_linrequiremnts_21u99v.html
• https://www.ibm.com/support/knowledgecenter/STSLR9_8.3.1/com.ibm.fs9200_831.doc/svc_linux_settings.html
44 © 2018 IBM Corporation February 16, 2024 IBM Services
Fibre Channel Host Attachment and DB/App Settings
AIX
Disk Related Parameters for AIX running MPIO:
• rw_timeout - Preferred setting for rw_timeout is 120
• Algorithm - Preferred setting is algorithm=shortest_queue
• Path health check mode (hcheck_mode) – Preferred setting is hcheck_mode=nonactive
• hcheck_interval - Preferred setting is hcheck_interval=240
• timeout_policy - Preferred setting is timeout_policy=fail_path ( if available on the device )
• Recommended to increase database timeout values such that they can withstand 3*rw_timeout value prior to taking any drastic action. If the database vendor
cannot allow 3*rw_timeout value, then at least allow 2*rw_timeout value, such that the lower level devices drivers get more than one chance to redrive the
I/O.
• Enable the fast fail and dynamic tracking attributes for hosts systems that run an AIX® 5.2 or later operating system.
• Refer to the links mentioned in the references for more details and the GTS best practice guidance for AIX MPIO configuration.
References:
• https://www.ibm.com/support/pages/node/630015
• https://w3-connections.ibm.com/wikis/home?lang=en-us#!/wiki/Global%20Server%20Management%20Distributed%20SL/page/Recommendations%20for%20Multipath%20Device%20Driver%20%26%20Settings
• https://www.ibm.com/support/pages/node/697363
• https://www.ibm.com/support/knowledgecenter/STSLR9_8.3.1/com.ibm.fs9200_831.doc/svc_aixconfigovrw_21uyy3.html
Windows
• Ensure Dual physical HBA’s are installed with redundant physical connection to both public SAN fabric switches.
• Ensure that you install the required software on your host, including the following items:
Operating system service packs and patches
Host bus adapters (HBAs)
HBA device drivers
Multipathing drivers
Clustered-system software
• Recommendation for is to use Microsoft Native MPIO with MS DSM.
• The default value for DiskTimeoutValue is 120. There is no requirement to reduce this value. If you need to reduce, ensure a minimum
value of 60 seconds is configured.
• The timeout value of 60 seconds will be sufficient for a Quorum race where I/O pause length = Short lease time + Quorum race + Hold
time + Pending IO resumption.
References:
• https://www.ibm.com/support/knowledgecenter/STSLR9_8.3.1/com.ibm.fs9200_831.doc/svc_FChostswindows_cover.html
• https://docs.microsoft.com/en-us/powershell/module/mpio/set-mpiosetting?view=win10-ps
VMware
• The Storage Array Type plug-in for IBM® volumes is set to VMW_SATP_ALUA.
• The path selection policy is set to RoundRobin.
• Set “RoundRobin” IOPS to 1 (not 1000) to evenly distribute I/Os across as many ports on the system as possible.
• Ensure the queue depth no deeper than 64.
• The number of paths per volume is recommended not to exceed 8. Ensure a minimum of 4 paths per volume.
• Do not use Datastores greater than 4TB.
• Ensure to leave the APD timeout value to a default of 140 secs.
• Disable auto removal of paths to a disk that is in PDL. Set “Disk.AutoremoveOnPDL” to 0.
References:
• https://www.ibm.com/support/knowledgecenter/STSLR9_8.3.1/com.ibm.fs9200_831.doc/svc_vmwconfigovrw_21lbjv.html
• https://docs.vmware.com/en/VMware-vSphere/5.5/com.vmware.vsphere.storage.doc/GUID-75214ECC-DC3B-4C06-95D4-221A8C03B94A.html
• Ensure that the queue depth total for all hosts that are connected to a single physical Fibre Channel port on the storage system does not
exceed 2048.
• If host I/O queue full failures regularly occur, consider reducing the queue depths or distributing the host zoning across more physical Fibre
Channel ports on the storage system.
Reference:
https://www.ibm.com/support/knowledgecenter/STSLR9_8.3.1/com.ibm.fs9200_831.doc/svc_FCqueuedepth.html
Reference(s):
https://www.ibm.com/support/knowledgecenter/STSLR9_8.3.1/com.ibm.fs9200_831.doc/svc_expandvolume.html
p
wa
wa
wa
wa
erS
erS
erS
erS
yp
yp
yp
yp
n/ H
n/ H
n/ H
n/ H
Hy t ap
p
ap
wa
wa
ti o
ti o
ti o
ti o
w
w
ca
ca
HoperS
ca
rS
ca
rS
HoperS
pl i
pl i
pl i
pe
pl i
pe
Fr st
Re st
Hy t
Fr st
Re st
Fr t
Re st
Fr st
Re st
Ho e
Ho e
ee
Ho e
s
s
e
e
Ho
Ho
Ho
Ho
Ho
Ho
Ho
Hy
Hy
1 2 3 4 1 2 3 4 1 2 3 4 1 2 3 4 1 2 3 4 1 2 3 4
1 1 1 1 2 2 2 2 3 3 3 3 1 1 1 1 2 2 2 2 3 3 3 3
FC Adapter1 FC Adapter2 FC Adapter3 FC Adapter1 FC Adapter2 FC Adapter3
CPU2 CPU1/CPU2 CPU1 CPU2 CPU1/CPU2 CPU1
p
wa
wa
wa
wa
erS
erS
erS
erS
yp
yp
yp
yp
n/ H
n/ H
n/ H
n/ H
Hy t ap
p
ap
wa
wa
ti o
ti o
ti o
ti o
w
w
ca
ca
HoperS
ca
rS
ca
rS
HoperS
pl i
pl i
pl i
pe
pl i
pe
Fr st
Re st
Hy t
Fr st
Re st
Fr t
Re st
Fr st
Re st
Ho e
Ho e
ee
Ho e
s
s
e
e
Ho
Ho
Ho
Ho
Ho
Ho
Ho
Hy
Hy
1 2 3 4 1 2 3 4 1 2 3 4 1 2 3 4 1 2 3 4 1 2 3 4
1 1 1 1 2 2 2 2 3 3 3 3 1 1 1 1 2 2 2 2 3 3 3 3
FC Adapter1 FC Adapter2 FC Adapter3 FC Adapter1 FC Adapter2 FC Adapter3
CPU2 CPU1/CPU2 CPU1 CPU2 CPU1/CPU2 CPU1
What happens if I write a highly compressible workload to an FCM? GTS Effective Capacity 364 TB
HS CV
Airbag
Even if you write a 10:1 compressible workload to an FCM, it will still be full Compression 2:1 331 TiB
when it reaches the maximum effective capacity. Any spare data space remaining
at this point will be used to improve the performance of the module and extend 80% Full Critical action required
70% Warning Alert
the wear. to stop new allocations
Usable Capacity 182 TB Threshold triggered