You are on page 1of 8

RAID [Redundant Array of Independent (Inexpensive) Disk]

---------------------------------------------------------RAID 0 Technolo y! Striping Data with No Data Protection. "erfor#ance! Highest $verhead! None %ini#u# &u#'er of Drives! 2 since striping Data (oss! Upon one drive failure )xa#ple! 5TB of usable space can be achieved through 5 dis". Advanta es! > High Perfor#ance !TB of

Disadvanta es! $uaranteed Data loss *ot +pare! Upon a drive failure% a hot spare can be invo"ed% but there will be no data to cop& over. Hot Spare is not a good option for this '()D t&pe. +upported! *lariion% S&##etri % S&##etri D'./s0 D+, -+eta B*./s or

)n '()D 1% the data is written 2 stripped across all of the dis"s. This is great for perfor#ance% but if one dis" fails% the data will be lost because since there is no protection of that data. -------------------------------------------------

RAID , Technolo y! +irroring and Duple ing "erfor#ance! Highest $verhead! 513 %ini#u# &u#'er of Drives! 2 Data (oss! ! Drive failure will cause no data loss. 2 drive failures% all the data is lost. )xa#ple! 5TB of usable space can be achieved through !1 dis". Advanta es! Highest Perfor#ance% 4ne of the safest. Disadvanta es! High 4verhead% (dditional overhead on the storage subs&ste#. Upon a drive failure it beco#es '()D 1. *ot +pare! ( Hot Spare can be invo"ed and data can be copied over fro# the surviving paired drive using Dis" cop&. +upported! *lariion% S&##etri % S&##etri D+, !TB of

The e act data is written to two dis"s at the sa#e ti#e. Upon a single drive failure% no data is lost% no degradation% perfor#ance or data integrit& issues. 4ne of the safest for#s of '()D% but with high overhead. )n the old da&s% all the S&##etri supported '()D ! and '()D S. Highl& reco##ended for high end business critical applications. The controller #ust be able to perfor# two concurrent separate 'eads per #irrored pair or two duplicate 5rites per #irrored pair. 4ne 5rite or two 'eads are possible per #irrored pair. Upon a drive failure onl& the failed dis" needs to be replaced.

RAID ,-0 Technolo y! +irroring and Striping Data "erfor#ance! High $verhead! 513 %ini#u# &u#'er of Drives! 6 Data (oss! Upon ! drive failure -+!0 device% no issues. 5ith #ultiple drive failures in the stripe -+!0 device% no issues. 5ith failure of both the +! and +2 data loss is certain. )xa#ple! 5TB of usable space can be achieved through !1 dis". !TB of

Advanta es! Si#ilar 7ault Tolerance to '()D 5% Because of striping high )24 is achievable. Disadvanta es! Upon a drive failure% it beco#es '()D 1. *ot +pare! Hot Spare is a good option with this '()D t&pe% since with a failure the data can be copied over fro# the surviving paired device. +upported! *lariion% S&##etri % S&##etri D+, '()D !81 is i#ple#ented as a #irrored arra& whose seg#ents are '()D 1 arra&s.

RAID . Technolo y! Striping Data with dedicated Parit& Drive. "erfor#ance! High $verhead! 993 4verhead with Parit& -in the e a#ple above0% #ore drives in 'aid 9 configuration will bring overhead down. %ini#u# &u#'er of Drives! 9 Data (oss! Upon ! drive failure% Parit& will be used to rebuild data. Two drive failures in the sa#e 'aid group will cause data loss. )xa#ple! 5TB of usable space would be achieved through : !TB dis". Advanta es! .er& high 'ead data transfer rate. .er& high 5rite data transfer rate. Dis" failure has an insignificant i#pact on throughput. ;ow ratio of <** -Parit&0 dis"s to data dis"s which converts to high efficienc&. Disadvanta es! Transaction rate will be e=ual to the single Spindle speed *ot +pare! ( Hot Spare can be configured and invo"ed upon a drive failure which can be built fro# parit& device. Upon drive replace#ent% hot spare can be used to rebuild the replaced drive. +upported! *lariion

RAID / Technolo y! Striping Data with Distributed Parit&% Bloc" )nterleaved Distributed Parit& "erfor#ance! +ediu#

$verhead! 213 in our e a#ple% with additional drives in the 'aid group &ou can substantiall& bring down the overhead. %ini#u# &u#'er of Drives! 9 Data (oss! 5ith one drive failure% no data loss% with #ultiple drive failures in the 'aid group data loss will occur. )xa#ple! 7or 5TB of usable space% we #ight need > ! TB drives

Advanta es! )t has the highest 'ead data transaction rate and with a #ediu# write data transaction rate. ( low ratio of <** -Parit&0 dis"s to data dis"s which converts to high efficienc& along with a good aggregate transfer rate. Disadvanta es! Dis" failure has #ediu# i#pact on throughput. )t also has #ost co#ple controller design. 4ften difficult to rebuild in the event of a dis" failure -as co#pared to '()D level !0 and individual bloc" data transfer rate sa#e as single dis". (s" the PS</s about '()D 5 issues and data loss? *ot +pare! Si#ilar to '()D 9% where a Hot Spare can be configured and invo"ed upon a drive failure which can be built fro# parit& device. Upon drive replace#ent% hot spare can be used to rebuild the replaced drive. +upported! *lariion% S&##etri D+, code @! '()D ;evel 5 also relies on parit& infor#ation to provide redundanc& and fault tolerance using independent data dis"s with distributed parit& bloc"s. <ach entire data bloc" is written onto a data dis"A parit& for bloc"s in the sa#e ran" is generated on 5rites% recorded in a distributed location and chec"ed on 'eads. This would classif& to be the #ost favorite '()D Technolog& used toda&.

RAID 0

Technolo y! Striping Data with Double Parit&% )ndependent Data Dis" with Double Parit& "erfor#ance! +ediu# $verhead! 2B3 in our e a#ple% with additional drives &ou can bring down the overhead. %ini#u# &u#'er of Drives! 6 Data (oss! 5ith one drive failure and two drive failures in the sa#e 'aid $roup no data loss. .er& reliable. )xa#ple! 7or 5 TB of usable space% we #ight need @ !TB drives

Advanta es! '()D > is essentiall& an e tension of '()D level 5 which allows for additional fault tolerance b& using a second independent distributed parit& sche#e -twoCdi#ensional parit&0. Data is striped on a bloc" level across a set of drives% Dust li"e in '()D 5% and a second set of parit& is calculated and written across all the drivesA '()D > provides for an e tre#el& high data fault tolerance and can sustain #ultiple si#ultaneous drive failures which t&picall& #a"es it a perfect solution for #ission critical applications. Disadvanta es! .er& poor 5rite perfor#ance in addition to re=uiring N82 drives to i#ple#ent because of twoCdi#ensional parit& sche#e. *ot +pare! Hot Spare can be invo"ed against a drive failure% built it fro# parit& or data drives and then upon drive replace#ent use that hot spare to build the replaced drive. +upported! *lariion 7lare 2>% 2B% S&##etri D+, *ode @2% @9 *lariion 7lare *ode 2> supports '()D >. )t is also being i#ple#ented with the @2 code on the S&##etri D+,. The si#plest e planation of '()D > is double the parit&. This allows a '()D > '()D $roups to be able to have two drive failures in the '()D $roup% while #aintaining access to the data.

RAID + (.-,) Technolo y! '()D S&##etri

"erfor#ance! 1 High $verhead! 253 %ini#u# &u#'er of Drives! 6 Data (oss! Upon two drive failures in the sa#e 'aid $roup )xa#ple! 7or 5 TB of usable space% B ! TB drives

Advanta es! High Perfor#ance on S&##etri <nviron#ent Disadvanta es! Proprietar& to <+*. '()D S can be i#ple#ented on S&##etri B111% 5111 and 9111 Series. Enown to have bac"end issues with director replace#ents% S*S) *hip replace#ents and bac"end D( replace#ents causing DU or offline procedures. *ot +pare! Hot Spare can be invo"ed against a failed drive% data can be built fro# the parit& or the data drives and upon a successful drive replace#ent% the hot spare can be used to rebuild the replaced drive. +upported! S&##etri B111% 5111% 9111. 5ith the D+, platfor# it is Dust called '()D -98!0 <+* S&##etri 2 D+, dis" arra&s use an alternate% proprietar& #ethod for parit& '()D that the& call '()DCS. Three Data Drives -,0 along with 4ne Parit& device. '()DCS is proprietar& to <+* but see#s to be si#ilar to '()DC5 with so#e perfor#ance enhance#ents as well as the enhance#ents that co#e fro# having a highCspeed dis" cache on the dis" arra&. The data protection feature is based on a Parit& '()D -98!0 volu#e configuration -three data volu#es to one parit& volu#e0. RAID (2-,) Technolo y! '()D S&##etri "erfor#ance! High $verhead! !2.53 %ini#u# &u#'er of Drives! B Data (oss! Upon two drive failures in the sa#e 'aid $roup

)xa#ple! 7or 5 TB of usable space% B get @ TB0

! TB drives -rather &ou will

Advanta es! High Perfor#ance on S&##etri <nviron#ent Disadvanta es! Proprietar& to <+*. (vailable onl& on S&##etri D+, Series. Enown to have a lot of bac"end issues with director replace#ents% bac"end D( replace#ents since &ou have to verif& the spindle locations. *ause of concern with DU. *ot +pare! Hot Spare can be invo"ed against a failed drive% data can be built fro# the parit& or the data drives and upon a successful drive replace#ent% the hot spare can be used to rebuild the replaced drive. +upported! 5ith the D+, platfor# it is Dust called '()D -@8!0. Not supported on the S&##s. <+* D+, dis" arra&s use an alternate% proprietar& #ethod for parit& '()D that is called '()D. Seven Data Drives -,0 along with 4ne Parit& device. '()D is proprietar& to <+* but see#s to be si#ilar to '()DCS or '()D5 with so#e perfor#ance enhance#ents as well as the enhance#ents that co#e fro# having a highCspeed dis" cache on the dis" arra&. The data protection feature is based on a Parit& '()D -@8!0 volu#e configuration -seven data volu#es to one parit& volu#e0.

You might also like