You are on page 1of 18

Dell Customer Communication - Confidential

==============
Training Plan
==============

 ISM : Information storage management


 Product : VNX ( File + Block )
 Hardware
 Software

+++++++++++++++
Day 1 :: ISM :
+++++++++++++++

Process part
Data /Big data
Data Center /SAN ( server , storage and switch/networking )
Cloud /Virtualization
File system structure
Drive structure
Drive types
++++++++++++++++++++++++++++++++++++++++++++++

Process Overview :

Mid-Range (USD storage produduct : VNX , VNXe and Unity


High-end or Enterprise Product : Isilion , symmetrix etc
Dell Customer Communication - Confidential

Channels: Dial Home , Vic/web , Chat


SR : Service request
Severity : S1 ( high impact , DU ( data unavailability ) , multilple faults etc )
S2 ,S3 : Low severity

Service cloud : Queue management : (24 * 7) >> to manage the newly logged in tickets /cases
>> Leads will manage the queue and assign the case.
Bin: Self assigned cases can be manage.

>> Different time zones ::


-----------------------------------
APJK : Asia Pacific Japan Korea
EMEA: Europe, Middle east, Africa
AMER: south and North America

S1 : SLA (service level agreement ) : 15min :: IR ( initial response via email and phone ) :: Needs
regular updates
S2 : SLA : 2-3 hrs
S3 : SLA 4-5 hrs

CSAT : Customer satisfaction


DSAT : Customer di-satisfaction
++++++++++++++++++++++++++++++++++
Dell Customer Communication - Confidential

Drive Types :

Electronic Drive : Extreme performance drives ::: ( IOPS : 3000-4000 IOPs) ::: ( capacity : 200 -
400 GB)
SSD ( solid state drive ) bootable device
Flash Drives ( SAS FLASH 2 , SAS FLASH 3 )
USB drives ( pendrives)

Mechanical Drive :::::

SAS ( serial attached scsi) ::::: ( IOPS = 300 -500 ) capacity ( 400GB -1500GB) >>
Performance drive
RPM – 15K ( more the rotation per minute less will be the seek and rotational disk latency and
higher will be the IOPS or processing time ). Eg. 400 IOPS
RPM - 10K >> 250 IOPS

NL-SAS ( Near line –serial attached scsi ) :::: (IOPS = 100 -300) == Capacity ( 800Gb – 2 TB -
4TB) >> Capacity drives
RPM -7.2K
IOPS : input /output operations
Dell Customer Communication - Confidential

+++++++++
Day 2:
+++++++++

RAID : Redundant array of inexpensive disk

RAID Types/Levels:

RAID 0 ( Stipping )
RAID 1 ( Mirroring )
RAID 10 ( Striped + Mirror )
RAID 5 >>> RAID width ( 2+1 , 3+1 , 4+1 , 8+1, 12+1 ) >> ( Distributed parity )

+1 >> parity space >> data can be reconstructed if single drive failure
9 drives >> 8+1 >> 1TB >> 9 TB ( 8 TB production + 1TB parity )
9 drives >> 8+1 >> 2TB >> 18TB ( 16 TB production + 2 TB parity )
80% ( production space )
RAID penalty : 4 ( 2 read operation + 2 write operation )

RAID 6 >> RAID width ( 2+2 , 4+2 , 8+2 etc . ) >> ( Dual distributed parity )
+ 2 >> parity space
Sustain two drive failure >> data can be reconstructed
Dell Customer Communication - Confidential

60% ( almost production space )


RAID penalty : 6 ( 3 read operation + 3 write operation )

Hotspare >>>>> spare drive >> 1 drive for every 30 drive by default ..can configure more than 1
as hotspare.
>>>>>> hotspare should be of same drive type and same or bigger drive size to invoke for
faulted drive.

DFRG : Double faulted raid Group ::::::::

Condition 1. ( No hotspare)
=============================
RAID 5 ( 4+1 )
Drive failure can sustain : 1

Single drive fails >> data can be reconstructed from parity .


Two drive fails >> data can not be reconstructed from parity >> may leads to DL ( data loss )

Condition 2. ( hotspare )
=========================

Rebuilding (REB) : Is the process of copying the data from faulted drive to hotspare drive
( faulted drive --- > HS drive )
RAID 5 ( 4+ 1) : 10 (11:25( HS )) 12 ( faulted):parity 13 14 ( Active )
Dell Customer Communication - Confidential

Drive failure can sustain : 2 ( 1 parity , 1 hotspare >> if drive rebuilding is completed )
3rd drive failure results in DFRG , may leads to DL .

Condition 3 . ( 2 hotspare ):
=============================
Drive failure can sustain : 3 ( 1 parity , 2 hotspare >> if drive rebuilding is completed )

Condition 4 ( 1 hotspare ):
============================
RG1: RAID 5 ( 4+1 ) : 10 11:HS25( REB :30%) 12:(failed) 13 14 25 ( HS ) >> DFRG ( double
faulted raid group ) as rebuilding is not
completed for drive 11 .

>> DFRG -- > DL (recover chances 70% )

>> steps taken to resolve the DFRG :

1. Re-seat the drive ( second faulted drive )


2. Re-seat the drive (multiple times )
3. Cool down ( freezer ) : 30min
4. Replacement of paddle card .
If still second faulted drive is not online then Declare as DL ::

Equalization ( EQZ ) : is the process of copying back the data from hostspare drive to newly
replaced drive . <-- HS
Dell Customer Communication - Confidential

RAID 5 : DFRG :may leads to DL


RAID 6 : TFRG :may leads to DL ( 3 drive failure )

RAID 6 ( 4+2 ) : 10 11:HS(25) 12:p1 13:p2 14 15 ( Active drives ) >> 25:HS


Increase drive failure limit from 2 to 3 with the help of hotspare . NO DFRG as all drive 11
completed with rebuilding with HS: 25 and for other
two drives data will be reconstructed from parity p1 and p2 resp.

RAID 6 ( 4+2 ) : 10 (11:HS(25)REB:50%) 12:p1 13:p2 14 15 ( Active drives ) >> 25:HS


TFRG : As 3 drives faulted at same time , rebuilding has not completed for drive 11.

Hotspare :
++++++++++++

HS : 30 ( 1/1.5/2TB : 15 RPM ) : size and RPM should match .

Hotspare policy : Every 30 drives : 1 HS default


35 drives : 2 HS
45 drives : 2 HS

Limitations of RG:
+++++++++++++++++++

1. Can not use different drive types in same RG


RG1 : RAID 5 : 3 SAS + 2 Nl_SAS >> not supported
RG2 : RAID 6 : NL_SAS/Flash
Dell Customer Communication - Confidential

2. Expansion is possible with similar types of drives

RG 1 : SAS : 4+1 >>> 4+1 ( NL-SAS )

3. Maximum 16 drives in one RG :

Pool :
++++++++

SAS Drive : 1 TB
RG1 ( RAID 5 ( 4+1 )) + RG2 ( RAID 6 (4+2) ) + RG3 (RAD 10 (3+3 ) == Pool == Total capacity 11
TB ( Total drives : 17 TB)

Pool : Container where the total storage space will be available . collection of disk
Homogeneous pool : Same types of drives : RG1 RAID 5 ( 4+1 ) + RG2 RAID 6 ( 4+ 2) + RG3 RAID
5 ( 8+1 ) >> SAS or NL_SAS or FLASH

Heterogeneous pool : Different types of drives : SAS RG4 + NL_SAS RG5 + FLASH (RG6) >>

Feature ( VP : virtual provisioning )


+++++++++++++++++++++++++++++++++++++
>> File system : 9TB

Heterogeneous Pool:
Pool 0: RG1 ( 4TB) + RG2 ( 4TB )+ RG3 (3TB) : ( SAS drive + NL_SAS + FLASH )

Homogeneous Pool:
Dell Customer Communication - Confidential

Pool 1 : SAS
Pool 2 : NL_SAS
Pool 3 : FLASH

VP ( virtual provisioning ) : Tiering policy : Technique by which data re-arrangement will take
place across the different set of drives to
achieve maximum efficiency . Only applicable for heterogeneous pool .

Start with Highest performance Tier >> Start with flash drive and then data will demoted to SAS
drives
Data demoted with lower performance tier >> from SAS to NL-SAS
Auto tier >>

Pool Utilization :
+++++++++++++++++++++

Total Pool capacity: 11 TB


Used Pool Capacity ( File systems , LUNs , datastores ) : 5 TB
Remaining pool capacity : 6TB

DAY 3:
==============

Storage Resources : >>>


++++++++++++++++++++++++
Dell Customer Communication - Confidential

<<>> Block resource ( lower or granular layer of space)


LUN ( logical unit Number )
-----------------------------

>> Basic form storage resource


>> RAW space /Block

Datastore
--------------

>> Formatted ( VMFS >> Vmware /windows)


VMware datastore:>> VMware Host
HyperV datastore :>> windows Host

File formatting for Different OS :


------------------------------------------
Linux : Ext2, Ext3
Unix : UFS
Windows : NTFS /FAT 32
VMWware : VMFS

File resource ( upper layer space ):: Tree structure


-----------------------------------------------------
Files /folders /directory >> File tree
<<>> File System
<<>>( backup copy ) Snapshots
Dell Customer Communication - Confidential

VNX : File + Block >> Unified


Clarrion : Block array
Celerra : File array

Provisioning :
---------------

>> Thick : What is promised is given to you


LUN : 2TB : >> 2TB

Pool 11TB : Thick LUN ( 2TB: ) >> available space = 9TB

>> Thin : Not provisioned how much space is ask for . Initial allocation will be very small space .
LUN : 2TB : >> 2000 GB ( space coming from pool is 200 GB)

Pool 11 TB : Thin LUN ( 2TB) : Initial Allocation : 100Gb >> available space pool : 10.9TB
Dell Customer Communication - Confidential

++++++++++++++++++++++
Day 4. Protocols :
+++++++++++++++++++++

Block Protocols :

>> iSCSI ( internet small computer system interface )


>> FC ( fibre channel )

File Protocol :
>> CIFS ( common internet file system )
>> NFS (Network file system )

Uniq Identfier / Initiator :


For iSCSI : IQN : iqn.1991-05.com.microsoft:vwduwip1-0010.corp.emc.com
For FC : WWN : 10:28:56:45:90:34:56:34:5c

++++++++++++++++++++
Day 5 : VNX Hardware
++++++++++++++++++++

VNX1 series Models : VNX : 5100 ,5300, 5500 ,5700, 7500


VNX2 series Model : VNX :5200, 5400, 5600, 5800, 8000
Dell Customer Communication - Confidential

VNX gateway : VG2 , VG8 , VG10

DPE : Disk processore enclsoure ( SP + Disk )


DAE : Disk array enclosure ( Disk ) >> LCC A and LCC B ( link control card )
SPS : Standby power supply
DME : Data mover enclosure
CS : Control Station
SPS <--> SP cabling : Sense cable
DPE ( SP ) <--> DAE : SAS cable

SP components : CPU Cores, RAM , SSD , Power supply ( FAN ) , IO module ( ports ), battery ,
Mamagement ports , SAS ports , Processor

++++++++++++++++++++++
Day 6 : VNX Block
++++++++++++++++++++++
VNX Gateway Models : VG2 , VG8 , VG10 etc >>

Backend from VNX or symmetrix :


File Side ( CS and datamover ) + block ( SP + DAE + drives )
VNX File Attached to VNX block ( SP ) OR symmetrix Block.

Command useful to know backend : nas_storage –l


>> APM0000011676 ( alphanumeric – VNX )
>> 000000142376427 ( numeric – symmetrix )
Dell Customer Communication - Confidential

Block Management login : ( GUI >> graphical user interface , CLI >> command line interface )
Remotely anywhere : CLI ( SP A )/ ( SP B ) >>https://Sp_Ip:9519 ( Port number :9519 )
Unisphere : GUI ( https://sp_IP )

Credentials : username/password >> (local user : sysadmin/sysadmin)

Management port on SP >> RJ45 >>


( SP A IP ) : 10.10.10.11 /hostname
( SP B IP ) : 10.10.10.12 /hostname

File management login : Control station (Runs management task ) >> eth3 --> external access
connected to Management switch ( network ) : 10.10.10.13 ( eg. CS Ip )
Eth0 >> DM2
Eth1 >> CS1
Eth2 >> DM3
Eth3 >> external connection.

CLI : Putty : CS IP : username/password : nasadmin/nasadmin


GUI : Unisphere : ( https://CS_ip ) : nasadmin/nasadmin

Test box CS Ip : 10.241.169.170 ( nasadmin/nasadmin )


Dell Customer Communication - Confidential

Ownership LUN:
++++++++++++++++++++

RAID Group IDs Default LUN owner Current Owner


Odd-numbered SP A ( trespass ) SP B
Even-numbered SP B SP B

LUN state :
+++++++++++++++

Enable : ENA ( online )


Degraded: DEG ( partially faulted but online )
broken : BRK ( offline )

Drive state :
+++++++++++++++++++

REB : rebuilding ( copying the data from faulted drive to hotspare )


EQZ : Equalizing ( copying the data from hotspare to newly replaced drive )
BV : Background verification ( internal process ) >> coherency error ( inconsistency of data )
>> manually /automatic
BZR : Background zeroing
FRN : Foreign state ( not detected )
Bound : In use or part of any pool of RG
Unbound : Drive is detected but not in use or not a part of any pool/RG .
PUP : Powering up state
MIS : Missing state
Dell Customer Communication - Confidential

Day 7 : Host Configuration


+++++++++++++++++++++++++
LUN provisioning to Host : >>

LUN masking : Autherization :: Granting access of Luns to Host

LUN mapping : Mapping VNX LUN ID with HOST ID :: ALU/LUN ID <<>>> HLU

Zoning : Secured channel on switch ( zone with switch port called hard zone , zone with
software WWN called soft zone )

Permanent Sparing in Two series : faulted drive is permanently replaced with Spare drive

Cache : Temporary storage for effictive Management of data and to improve performance .

From Unisphere:
Disable write cache: System --> Under System Management click on System Properties --> Go
to SP Cache --> uncheck write cache (disable)

How to enable write cache?


From Unisphere:
Enable write cache: System --> Under System Management click on System Properties --> Go
to SP Cache --> check write cache (enable)
Dell Customer Communication - Confidential

Vault drives : ( First Four Drive ) : vaulting


Vault Drive : DPE drives 0.0.0, 0.0.1 , 0.0.2 , 0.0.3
Incase of power failure data will de-stage from cache into vault drives, its called cache vaulting .
If data is not able to de-stage after power failure then it leads to cache corruption called cache
dirty .Entire DU to system .

Vaults LUN : configurational data .


++++++++++++++++++++++++++
Remember that the vault drives are divided into two sections - the OS part that is hidden from
view and is the part that contains the OS, the Flare information and a bunch of other stuff. This
part of the drives is always in use by the OS - updating the location information and other
things.
The second part of the drives, if the drive is large enough in capacity, can be used to create a
User raid group and User LUNs. 2 In the VNX there are four vault drives so you could create a
2+R10 or a 3+1 R5 if you wanted.
EMC recommends that you do not use user space of the vault drives for any application what
will be using the disks a lot.
When a Vault drive fails how that is handled by the OS is different depending on if you have
User Raid Group and LUNs - if you have User LUNs, the disk will get replaced by the Hot Spare,
but it only replicates the User LUN space and not the OS space. That allows the User LUNs to
continue to work, but the OS space is still faulted and that can disable Write cache until the
failed drive is physically replaced.

vault LUNs created from vault drives carries configuration info. from file and block O.E.

FLARE : block O.E ( Fiber logic array in runtime environment)


>> SP A OS image ( drive 0 and 2 )
>> SP B OS image ( drive 1 and 3 )
PSM
Internal databases
Dell Customer Communication - Confidential

DART : File O.E ( Data access in real time ) >> Configuration will reside in control LUNs.

You might also like