Professional Documents
Culture Documents
Solutions
1
Huawei Converged Storage
Architecture and Technical Overview
1 Introduction to Converged Storage Products
3
Huawei Converged Storage Overview
OceanStor
OceanStor 5800 V5
OceanStor
OceanStor 5600 V5
OceanStor 5500 V5
2200/2600 V3 5300 V5
4
Multi-Level Convergence Makes Core Services More Agile
HyperMetro
Dorado
V5 V3
NAS+SAN SAN NAS SAN NAS
Multiple storage SAN and NAS SSD and HDD Multiple storage A-A for SAN and
types resource pools NAS
Pooling of heterogeneous Gateway-free converged data
Interconnection between Support for multiple types of HDDs and SSDs converged to
storage resources DR solution
different types, levels, and services meet the performance
Unified management and Smooth upgrade to 3DC
generations of flash storage Industry-leading performance requirements of complex
automated service orchestration
and functions services
Multi-level convergence
99.9999% service availability, satisfying complex service requirements
5
1 Introduction to Converged Storage Products
6
OceanStor Converged Storage V5 Architecture: Comparison
Converged and parallel file and block SAN over NAS file system SAN over NAS file system
services architecture architecture
RAID Manager
RAID DP / RAID4+
Storage pool RAID 2.0+
Storage subsystems
SAN
• Converged block and file services • Converged block and file storage
• Parallel processing of file system • WAFL-based architecture • Standalone NAS gateway
services and block services in a • Unified file & block manager • NAS storage pool consisting of
storage pool • Physical RAID group one or more LUNs mapped
• Storage pool based on RAID 2.0+ from SAN storage
7
OceanStor V5 Software Architecture Overview
8
Convergence of SAN and NAS: How to Implement
Disk Domain Storage Pool LUN & FS Protocol
Grain Level-1
(Thin & iSCSI/FC
CKG Level-2 cache Extent Dedupe &
Tiered cache
(RAID group) (Stripe cache) (Tiering unit) Compression) Thick LUN (LUN & FS cache)
SSD
Level-1 cache
Block-level Tiered
tiering Thin LUN
Level-2 cache
Tier0
SAS Not tiered
Thin LUN
Tier1
NL-SAS
Directly divided Grain File system
from CKG
Level-1
NFS share and CIFS share
cache
Tier2 File-level tiering
OceanStor converged storage architecture minimizes the I/O paths of SAN and NAS, and provides optimal performance.
9
OceanStor V5: Reliable Scale-out Storage
CKG CKG
App App
Active-Active
SSD HDD
10
Key Technology: RAID 2.0+ Architecture
Hot Hot
spare spare
11
More Reliable System with 20-Fold Faster Data Reconstruction
Reconstruction principle of Reconstruction principle With RAID 2.0+, data reconstruction time
traditional RAID of RAID 2.0+ plummets from 10 hours to 30 minutes
10
6
10 hours
4
2 30
minutes
0
Traditional Huawei's quick
technology recovery technology
12
SmartMatrix 3.0 Overview
13
Key Technology: SmartMatrix 3.0 Front-End Interconnect I/O Module (FIM)
14
Key Technology: SmartMatrix 3.0 Persistent Cache
A A* C C* A C C* A C
B* B D* D B D* D B D
A* B* C* A*
D* B*
A B C D A B C D A B C D
If controller A fails, controller B takes over its cache, and the cache of controller B (including that of controller A) is
mirrored to controller C or D. If controller D is also faulty, controller B or C mirrors its cache.
If a controller fails, services are switched rapidly to the mirror controller, and the mirror relationship between it and
other controllers is re-established, so each controller has a cache mirror. In this way, write-back (instead of write-
through) for service requests is ensured. This guarantees the performance after the controller failure and ensures
system reliability, because data written into the cache has mirror redundancy.
15
Key Technology: Load Balancing of SmartMatrix 3.0 Persistent Cache
Controller A Controller B Controller A Controller B
C1 D1' D1 C1'
C1 D1' D1 C1' C2 A1' D2 B2'
C2 A2' D2 B2' C3 B3' D3 A2'
C3 B3' D3 A3' A2 B1' A3 C2'
16
Key Technology: HyperMetro (Block & File)
Site A Site B
Working Principles
One device
Host application
Gateway-free, one device uses HyperMetro to support both active-active files
cluster (shared
volumes and databases.
mounted to One quorum system
HyperMetro file SAN and NAS share one quorum site. Services are provided by the same site
systems) in the event of link failures, ensuring data consistency.
One network
Storage network
between storage
The heartbeat, configuration, and physical links between two sites are
arrays and hosts integrated into one link. One network supports both SAN and NAS
IP&FC IP&FC transmission.
SAN
Real-time data
NAS SAN mirroring SAN NAS
Dual-write heartbeats
and configurations
Highlights
Active-Active, RPO = 0, RTO ≈ 0
Requires no gateway devices, simplifying networks, saving costs, and
FC/IP eliminating gateway-caused latency.
Production storage Production storage Supports two quorum servers, improving reliability
Supports flexible combination of high-end, mid-range, and entry-level
IP IP storage arrays for active-active solutions, saving investment.
Supports smooth upgrade from active-active or active-passive solutions to
3DC solutions without service interruption.
Flexibly supports 10GE or FC networks for intra-city interconnection and IP
networks for quorum links.
Quorum site
17
1 Introduction to Converged Storage Products
18
OceanStor Converged Storage Features
SAN Features
iSCSI FC
19
OceanStor Converged Storage Features
NAS Features
20
SmartMulti-Tenant Architecture: Network, Protocol, and Resource Virtualization
Protocol virtualization
vStore vStore
AD Client
Client
Client
Client
Client LDAP/NIS
Client
Lock manager
FS 0 FS 1 FS 2 FS 3 VFS
21
HyperMetro for NAS
Site A Site B
Working Principles
FS FS FS FS High-availability synchronous mirror at a file-system level: When data is
written to the primary file system, it will be synchronously replicated to the
secondary file system. If the primary site or file system fails, the secondary site
HyperMetro vStore pairs:
(vStore1 vStrore1') or file system will automatically take over services without any data loss or
IP/FC IP/FC
(vStore2 vStrore2') application disruption.
HyperMetro pairs:
FS1 FS1' (FS1 FS1')
vStore1 vStore1' (FS2 FS2')
FS2
Data and
FS2'
(FS3 FS3') Highlights
FS3 configuration FS3' (FS4 FS4')
vStore2 sync
vStore2'
Gateway-free deployment
FS4 FS4' 1 network type between sites
FS5 FS5' 2 components required for smooth upgrade
vStore3 vStore3'
FS6 FS6' 3 automatic fault recovery scenarios
4x scalability
5x switching speed
IP IP
Quorum server
22
HyperMetro Architecture
LADP, AD, and NIS servers
Client
File system
FS 0 FS 1 FS 2 FS' 0 FS' 1 FS' 2
23
Service and Access Processes in the Normal State
1-7: Configurations made by the administrator on
NAS client Admin
vStore A are synchronized to vStore A' in real
10.10.10.1 time, such as the quota, qtree, NFS service,
11 18 CIFS service, security strategy, user and user
1 7
10.10.10.10 10 8 10.10.10.10
group, user mapping, share, share permission,
DNS, AD domain, LDAP, and NIS.
2
NAS service CFG sync
4
CFG sync NAS service If a failure occurs, the changed configurations
12 17 2 3 5 6 are saved in the CCDB log and the vStore pair
status is set to "to be synchronized". After the
File system CCDB CCDB File system
link is recovered, the configurations in the CCDB
13 16 vStore A vStore A'
16
14 14 14 log are automatically synchronized to vStore A'.
Object set Data sync Data sync Object set
14 15
15 15 15 14 15 8-18: When a NAS share is mounted to the
Concurrent write
9
Cache Cache client, the storage system obtains the access
permission of the share path based on the client
IP address. If the network group or host name
Storage pool Storage pool
has the permission, the client obtains the handle
of the shared directory. If a user writes data into
a file, the NAS service processes the request
Storage system A Storage system B
and converts it into a read/write request of the
file system. If it is a write request, the data
LADP/NIS
synchronization module writes the data to the
server caches of both sites simultaneously, and then
returns the execution result to the client.
24
Service and Access Processes During a Failover
10.10.10.1
1 8
10
10.10.10.10 10.10.10.10 1-8: When vStore A is faulty, vStore A'
13
detects the faulty pair status and
NAS service CFG sync CFG sync NAS service applies for arbitration from the quorum
11 12 2 7 server. After obtaining the arbitration,
File system CCDB CCDB File system vStore A' activates the file system, NAS
vStore A vStore A' 3 6 service, and LIF status. NAS service
Object set
14 Data
CCDB log Object set configuration differences are recorded
synchronization 4 5 in the CCDB log, and data differences
Cache DCL
9 Cache are recorded in the data change log
9 (DCL). In this manner, vStore A can
Storage pool Storage pool synchronize incremental configurations
and data upon recovery.
Storage system A Storage system B
The CCDB log and DCL are configured
with power failure protection and have
high performance.
LADP/NIS server
25
NFS Lock Failover Process
Synchronize a client's IP address pair Notify the client The client reclaims the lock
Mount1: Mount1:
10.10.10.11:/fs1 Mount1:
10.10.10.11:/fs1
10.10.10.11:/fs1
10.10.10.1 10.10.10.1
10.10.10.1
Notify Reclaim
10.10.10.11 10.10.10.11
10.10.10.11 (inactive) 10.10.10.11 10.10.10.11
(inactive) 10.10.10.11
Back up
Configuration client info Configuration Read
synchronization synchronization configuration
CCDB
CCDB CCDB
1. HyperMetro backs up a 1. The NAS storage reads the list 1. The client sends a lock
client's IP address pair to of IP address pairs from the reclaiming command to the
remote storage. CCDB. storage.
2. The NAS storage sends 2. The storage recovers byte-
NOTIFY packages to all clients range locks.
to reclaim locks.
26
NAS HyperMetro: FastWrite
RTT-2
8. Status good
27
HyperSnap
The host
modifies data D
Before snapshot During snapshot Modify data after a
creation creation snapshot is created
Active Snapshot Active Snapshot Active
volume mapping table volume mapping table Volume Copy-on-write (COW)
Used by LUNs of OceanStor
A B C D A B C D A B C D1 D V5 storage
New
data Modified
data
Data block D is copied and the mapping table is modified.
A B C D A B C D A B C D B D1 E1 E2
Deleted Modified New
data data data
28
HyperVault
2.1 Working Principles (1)
The initial backup of HyperVault is
a full backup, and subsequent
backups are incremental backups.
based on file systems, the backups
are completely transparent to hosts
and applications.
Each copy at the backup file
system contains full service data,
not only the incremental data.
The data at the backup file system
is stored in the original format and
is readable after the backup is
complete.
29
HyperVault
30
DR Star (SAN)
I/O process:
DC 1
1. The host delivers I/Os to the primary LUN-A.
LUN-A (Ta) 2. The primary site dual-writes the I/Os to the secondary LUN-B.
Asynchronous replication (standby)
3. A write success is returned to the host.
4 4. Asynchronous replication starts and triggers LUN-A to activate the time slice
1 Ta+1. New data written to LUN-A is stored in this time slice, and the Ta slice is
LUN-A (Ta+1) DC 3 used as the data source for the standby asynchronous replication.
3
5. LUN-B activates a new time slice Tb+1, and the new data is stored in this time
LUN-C (Tc+1) slice. LUN-C activates a new time slice Tc+1 as the target of asynchronous
replication. Tc is the protection point of asynchronous replication rollback.
Active-active 2 5 6. LUN-B (Tb) is the data source for asynchronous replication to LUN-C (Tc+1).
Standby
DC 2 6 host
Data in DC1 and DC2 is synchronous. After the data is copied from Tb to Tc+1,
LUN-C (Tc)
the data in Ta is also copied to Tc+1. This process is equivalent to the
LUN-B (Tb+1) asynchronous replication between DC1 and DC3. If DC2 is faulty, DC1 and DC3
5 are switched to asynchronous replication. Incremental data is replicated from Ta
Asynchronous replication to DC3.
Standby
LUN-B (Tb)
host
Compared with the common 3DC solution:
1. There is a replication relationship between every two sites. Only one of the
Item Huawei H** E**
two asynchronous replication relationships has I/O replication services and
Active-Active + the other one is in the standby state.
asynchronous remote Supported Supported Not supported 2. If the working asynchronous replication link is faulty or one of the active-
replication active sites is switched over, the working link is switched to the standby link.
Synchronous remote Then, incremental synchronization can be implemented.
replication + asynchronous Supported Not supported Supported 3. You only need to configure DR Star at one site.
remote replication 4. The DR Star supports active-active + asynchronous + standby and
synchronous + asynchronous + standby networking modes. The
Configured at one site Supported Not supported Supported
asynchronous + asynchronous + standby networking mode is not supported.
31
SmartTier (Intelligent Tiering)
LUN
Tier0: SSD Tier1: SAS Tier2: NL-SAS Relocates data based on the rank
Data relocation
and relocation policy.
ROOT
Dir Dir
File system
Indicates the user-defined file write policy and
relocation policy.
The supported attributes include the file
File policy
size/name/type/atime/ctime/crtime/mtime.
File
Scans the list of files to be relocated based on
File distribution analysis the file policy.
Tier0: SSD Tier1: SAS/NL-SAS File relocation Relocates files based on policy.
32
SmartTier for NAS (Intelligent File Tiering)
File system
Highlights
Tiering policy
33
SmartTier File Relocation Principles
SmartTier policy
Relocation period
• Specifies the start time.
Add files to the background • Specifies the running duration.
relocation task. • Can be Paused.
34
SmartTier for NAS and Background Deduplication & Compression
Configure SmartTier to improve performance and save space:
Enable SmartTier for the file system and configure the automatic relocation mode where all data is written into the performance tier
(SSD tier).
Set the SmartTier relocation time from 22:00 to 05:00.
In SmartTier, enable deduplication and compression during relocation.
Performance tier
(SSD)
Create a file system. New data is Data is deduplicated and Deduplication New data is
written to SSDs. compressed when and compression written to SSDs.
relocated to HDDs. are complete.
35
THANK YOU