You are on page 1of 31

1. Evolution of various Storage Technologies.

Evolution of Storage Technology:


Storage or storage devices are defined as the devices for storing data that is created by individuals or businesses so that it can be easily accessible for further processing. Devices such as memory in a cell phone or digital camera, DVDs, CD-ROMs, and hard disks in personal computers are examples of storage devices. Businesses have several options available for storing data including internal hard disks, external disk arrays and tapes. Organizations had centralized computers (mainframe) and information storage devices (tape reels and disk packs) in their data center. The proliferation of departmental servers in an enterprise resulted in unprotected, unmanaged, fragmented islands of information and increased operating cost. To overcome these challenges, storage technology evolved from non-intelligent internal storage to intelligent networked storage. Evolution of storage technologies are as follows:

v`u Redundant

Array of Independent Disks (RAID): This technology was developed to address the cost, performance, and availability requirements of data. It continues to evolve today and is used in all storage architectures such as DAS, SAN, and so on. vauDirect-attached storage (DAS): This type of storage connects directly to a server (host) or a group of servers in a cluster. Storage can be either internal or external to the server. External DAS alleviated the challenges of limited internal storage capacity. vbu Storage area network (SAN): This is a dedicated, high-performance Fiber Channel (FC) network to facilitate block-level communication between servers and storage. Storage is partitioned and assigned to a server for accessing its data. SAN offers scalability, availability, performance, and cost benefits compared to DAS. vcu Network-attached storage (NAS): This is dedicated storage for file serving applications. Unlike a SAN, it connects to an existing communication network (LAN) and provides file access to heterogeneous clients. Because it is purposely built for providing storage to file server applications, it offers higher scalability, availability, performance, and cost benefits compared to general purpose file servers. vdu Internet Protocol SAN (IP-SAN): One of the latest evolutions in storage architecture, IP-SAN is a convergence of technologies used in SAN and NAS. IPSAN provides block-level communication across a local or wide area network (LAN or WAN), resulting in greater consolidation and availability of data.

2. Study of information life cycle management.

Data, whether structured or unstructured, does not fulfill any purpose for individuals or businesses unless it is presented in a meaningful form. Businesses need to analyze data for it to be of value. Information is the intelligence and knowledge derived from data.

KNOWLEDGE

INFORMATION

DATA

Key challenges for managing information are as follows:


v`u Exploding

digital universe. vauIncreasing dependency on information. vbu Changing value of information. The information lifecycle is the change in the value of information over time. When data is first created, it often has the highest value and is used frequently. As data ages, it is accessed less frequently and is of less value to the organization. Understanding the information lifecycle helps to deploy appropriate storage infrastructure, according to the changing value of information. For example, in a sales order application, the value of the information changes from the time the order is placed until the time that the warranty becomes void.

PROJECT

New Order

Process Order

Deliver Order

Warranty Claim

VALUE

Fulfilled order

Aged data

Warranty Voided

Create

Access

Migrate

Archive

Dispose

The value of the information is highest when a company receives a new sales order and processes it to deliver the product. After order fulfillment, the customer or order data need not be available for real-time access. The company can transfer this data to less expensive secondary storage with lower accessibility and availability requirements unless or until a warranty claim or another event triggers its need. After the warranty becomes void, the company can archive or dispose of data to create space for other high-value information.

Information Lifecycle Management

Centralized information storage and processing

Information Creator

Information User

Information lifecycle management (ILM) is a proactive strategy that enables an

IT organization to effectively manage the data throughout its lifecycle, based on predefined business policies. An ILM strategy should include the following characteristics:
v`u Business-centric:

It should be integrated with key processes, applications, and initiatives of the business to meet both current and future growth in information. vauCentrally managed: All the information assets of a business should be under the purview of the ILM strategy. vbu Policy-based: The implementation of ILM should not be restricted to a few departments. ILM should be implemented as a policy and encompass all business applications, processes, and resources. vcu Heterogeneous: An ILM strategy should take into account all types of storage platforms and operating systems. vdu Optimized: Because the value of information varies, an ILM strategy should consider the different storage requirements and allocate storage resources based on the informations value to the business.

ILM Implementation
The process of developing an ILM strategy includes four activities classifying, implementing, managing, and organizing: Classifying data and applications on the basis of business rules and policies to enable differentiated treatment of information. Implementing policies by using information management tools, starting from the creation of data and ending with its disposal. Managing the environment by using integrated tools to reduce operational complexity. Organizing storage resources in tiers to align the resources with data classes, and storing information in the right type of infrastructure based on the informations current value. Implementing ILM across an enterprise is an ongoing process. Figure illustrates a three-step road map to enterprise-wide ILM.

ILM Benefits
Implementing an ILM strategy has the following key benefits that directly address the challenges of information management:
v`u Improved

utilization by using tiered storage platforms and increased visibility of all enterprise information.

vauSimplified

management by integrating process steps and interfaces with individual tools and by increasing automation. vbu A wider range of options for backup, and recovery to balance the need for business continuity. vcu Maintaining compliance by knowing what data needs to be protected for what length of time. vdu Lower Total Cost of Ownership (TCO) by aligning the infrastructure and management costs with information value. As a result, resources are not wasted, and complexity is not introduced by managing low-value data at the expense of high-value data.

3. Study of disk physical structure components, properties and performance.


Physical disk structure components:
Data on the disk is recorded on tracks, which are concentric rings on the platter around the spindle, as shown in fig. The tracks are numbered, starting from zero, from the outer edge of the platter. The number of tracks per inch (TPI) on the platter (or the track density) measures how tightly the tracks are packed on a platter. Each track is divided into smaller units called sectors. A sector is the smallest, individually addressable unit of storage. The track and sector structure is written on the platter by the drive manufacturer using a formatting operation. The number of sectors per track varies according to the specific drive. The first personal computer disks had 17 sectors per track. Recent disks have a much larger number of sectors on a single track. There can be thousands of tracks on a platter, depending on the physical dimensions and recording density of the platter. Typically, a sector holds 512 bytes of user data, although some disks can be formatted with larger sector sizes. In addition to user data, a sector also stores other information, such as sector number, head number or platter number, and track number. This information helps the controller to locate the data on the drive, but storing this information consumes space on the disk. Consequently, there is a difference between the capacity of an unformatted disk and a formatted one. Drive manufacturers generally advertise the unformatted capacity for example, a disk advertised as being 500GB will only hold 465.7GB of user data, and the remaining 34.3GB is used for metadata. A cylinder is the set of identical tracks on both surfaces of each drive platter. The location of drive heads is referred to by cylinder number, not by track number. 1. Platter
A typical HDD consists of one or more flat circular disks called platters. The data is recorded on these platters in binary codes (0s and 1s). The set of rotating platters is sealed in a case, called a Head Disk Assembly (HDA). A platter is a rigid, round disk coated with magnetic material on both surfaces (top and bottom). The data is encoded by polarizing the magnetic area, or domains, of the disk surface. Data can be written to or read from both surfaces of the platter. The number of platters and the storage capacity of each platter determine the total capacity of the drive.

2. Spindle A spindle connects all the platters, as shown in Figure 2-3, and is connected to a motor. The motor of the spindle rotates with a constant speed. The disk platter spins at a speed of several thousands of revolutions per minute (rpm). Disk drives have spindle speeds of 7,200 rpm, 10,000 rpm, or 15,000 rpm. Disks used on current storage systems have a platter diameter of 3.5 (90mm). When the platter spins at 15,000 rpm, the outer edge is moving at around 25 percent of the speed of sound. The speed of the platter is increasing with improvements in technology, although the extent to which it can beimproved is limited. 3. Read/Write Head Read/Write (R/W) heads, shown in Figure 2-4, read and write data from or to a platter. Drives have two R/W heads per platter, one for each surface of the platter. The R/W head changes the magnetic polarization on the surface of the platter when writing data. While reading data, this head detects magnetic polarization on the surface of the platter. During reads and writes, the R/W head senses the magnetic polarization and never touches the surface of the platter. When the spindle is rotating, there is a microscopic air gap between the R/W heads and the platters, known as the head flying height. This air gap is removed when the spindle stops rotating and the R/W head rests on a special area on the platter near the spindle. This area is called the landing zone. The landing zone is coated with a lubricant to reduce friction between the head and the platter. The logic on the disk drive ensures that heads are moved to the landing zone before they touch the surface. If the drive malfunctions and the R/W head accidentally touches the surface of the platter outside the landing zone, a head crash occurs. In a head crash, the magnetic coating on the platter is scratched and may cause damage to the R/W head. A head crash generally results in data loss. 4. Actuator Arm Assembly The R/W heads are mounted on the actuator arm assembly which positions the R/W head at the location on the platter where the data needs to be written or read. The R/W heads for all platters on a drive are attached to one actuator arm assembly and move across the platters simultaneously. Note that there are two R/W heads per platter, one for each surface.

5. Controller The controller (see Figure 2-2 [b]) is a printed circuit board, mounted at the bottom of a disk drive. It consists of a microprocessor, internal memory, circuitry, and firmware. The firmware controls power to the spindle motor and the speed of the motor. It also manages communication between the drive and the host. In addition, it controls the R/W operations by moving the actuator arm and switching between different R/W heads, and performs the optimization of data access.

(6) Physical Disk Structure Data on the disk is recorded on tracks, which are concentric rings on the platter around the spindle, as shown in Figure. The tracks are numbered, starting from zero, from the outer edge of the platter. The number of tracks per inch (TPI) on the platter (or the track density) measures how tightly the tracks are packed on a platter. Each track is divided into smaller units called sectors. A sector is the smallest, individually addressable unit of storage. The track and sector structure is written on the platter by the drive manufacturer using a formatting operation. The number of sectors per track varies according to the specific drive. The first personal computer disks had 17 sectors per track. Recent disks have much larger number of sectors on a single track. There can be thousands of tracks on a platter, depending on the physical

dimensions and recording density of the platter. Typically, a sector holds 512 bytes of user data, although some disks can be formatted with larger sector sizes. In addition to user data, a sector also stores other information, such as sector number, head number or platter number, and track number. This information helps the controller to locate the data on the drive, but storing this information consumes space on the disk. Consequently, there is a difference between the capacity of an unformatted disk and a formatted one. Drive manufacturers generally advertise the unformatted capacity for example, a disk advertised as being 500GB will only hold 465.7GB of user data, and the remaining 34.3GB is used for metadata. A cylinder is the set of identical tracks on both surfaces of each drive platter. The location of drive heads is referred to by cylinder number, not by track number.

Properties
1. Zoned Bit Recording Because the platters are made of concentric tracks, the outer tracks can hold more data than the inner tracks, because the outer tracks are physically longer than the inner tracks, as shown in Figure 2-6 (a). On older disk drives, the outer tracks had the same number of sectors as the inner tracks, so data density was low on the outer tracks. This was an inefficient use of available space. Zone bit recording utilizes the disk efficiently. As shown in Figure 2-6 (b), this mechanism groups tracks into zones based on their distance from the center of the disk. The zones are numbered, with the outermost zone being zone 0. An appropriate number of sectors per track are assigned to each zone, so a zone near the center of the platter has fewer sectors per track than a zone on the outer edge. However, tracks within a particular zone have the same number of sectors. 2. Logical Block Addressing Earlier drives used physical addresses consisting of the cylinder, head, and sector (CHS) number to refer to specific locations on the disk, as shown in figure, and the host operating system had to be aware of the geometry of each disk being used. Logical block addressing (LBA), shown in Figure , simplifies addressing by using a linear address to access physical blocks of data. The disk controller translates LBA to a CHS address, and the host only needs to know the size of the disk drive in terms of the number of blocks. The logical blocks are mapped to physical sectors on a 1:1 basis. In Figure, the drive shows eight sectors per track, eight heads, and four cylinders. This means a total of 8 8 4 = 256 blocks, so the block number ranges from 0 to 255. Each block has its own unique address. Assuming that the sector holds 512 bytes, a 500 GB drive with a formatted capacity of 465.7 GB will have in excess of 976,000,000 blocks.

3. Disk capacity The maximum number of bits that can be recorded ye a disk is known as its maximum capacity, or simply capacity. Disk capacity is determined by the following technology factor: 1. Recoding density(bits/in): The number of bits that can be squeezed into a 1inch segment of a track. 2. Track density (tracks/in): The number of tracks that can be squeezed into a 1-inch segment of the radius extending from the center of the platter. 3. Areal density(bits/in2): The product of the recording density and the track density. Disk manufacturers work tirelessly to increase areal density(and thus capacity), and this is doubling every few years. The capacity of a disk is given by the following formula: Disk Capacity= (#bytes/sector)*(average#sectors/track)*(#tracks/surface) (#surfaces/platter)*(#platters/disk)

Performance
A disk drive is an electromechanical device that governs the overall performance of the storage system environment. The various factors that affect the performance of disk drives are discussed in this section.

Disk Service Time


Disk service time is the time taken by a disk to complete an I/O request. Components that contribute to service time on a disk drive are seek time, rotational latency, and data transfer rate. 1. Seek Time
The seek time (also called access time) describes the time taken to position the R/W heads across the platter with a radial movement (moving along the radius of the platter). In other words, it is the time taken to reposition and settle the arm and the head over the correct track. The lower the seek time, the faster the I/O operation. Disk vendors publish the following seek time specifications:

Full Stroke: The time taken by the R/W head to move across the entire width of the disk, from the innermost track to the outermost track.

Average: The average time taken by the R/W head to move from one random track to another, normally listed as the time for one-third of a full stroke. Track-to-Track: The time taken by the R/W head to move between adjacent tracks. Each of these specifications is measured in milliseconds. The average seek time on a modern disk is typically in the range of 3 to 15 milliseconds. Seek time has more impact on the read operation of random tracks rather than adjacent tracks. To minimize the seek time, data can be written to only a subset of the available cylinders. This results in lower usable capacity than the actual capacity of the drive. For example, a 500 GB disk drive is set up to use only the first 40 percent of the cylinders and is effectively treated as a 200 GB drive. This is known as short-stroking the drive. 2. Rotational Latency To access data, the actuator arm moves the R/W head over the platter to a particular track while the platter spins to position the requested sector under the R/W head. The time taken by the platter to rotate and position the data under the R/W head is called rotational latency. This latency depends on the rotation speed of the spindle and is measured in milliseconds. The average rotational latency is one-half of the time taken for a full rotation. Similar to the seek time, rotational latency has more impact on the reading/writing of random sectors on the disk than on the same operations on adjacent sectors. Average rotational latency is around 5.5 ms for a 5,400-rpm drive, and around 2.0 ms for a 15,000-rpm drive. 3. Data Transfer Rate The data transfer rate (also called transfer rate) refers to the average amount of data per unit time that the drive can deliver to the HBA. It is important to first understand the process of read and write operations in order to calculate data transfer rates. In a read operation, the data first moves from disk platters to R/W heads, and then it moves to the drives internal buffer. Finally, data moves from the buffer through the interface to the host HBA. In a write operation, the data moves from the HBA to the internal buffer of the disk drive through the drives interface. The data then moves from the buffer to the R/W heads. Finally, it moves from the R/W heads to the platters. The data transfer rates during the R/W operations are measured in terms of internal and external transfer rates, Internal transfer rate is the speed at which data moves from a single track of a platters surface to internal buffer (cache) of the disk. Internal transfer rate takes into account factors such as the seek time. External transfer rate is the rate at which data can be moved through the interface to the HBA. External transfer rate is generally the advertised speed of the interface, such as 133 MB/s for ATA. The sustained external transfer rate is lower than the interface speed.

4. Study of DAS, NAS and SAN.


DAS:
Direct-Attached Storage (DAS) is an architecture where storage connects directly to servers. Applications access data from DAS using block-level access protocols. The internal HDD of a host, tape libraries, and directly connected external HDD packs are some examples of DAS. DAS is classified as internal or external, based on the location of the storage device with respect to the host. 1. Internal DAS In internal DAS architectures, the storage device is internally connected to the host by a serial or parallel bus. The physical bus has distance limitations and can only be sustained over a shorter distance for high-speed connectivity. In addition, most internal buses can support only a limited number of devices, and they occupy a large amount of space inside the host, making maintenance of other components difficult. 2. External DAS In external DAS architectures, the server connects directly to the external storage device (see Figure 5-1). In most cases, communication between the host and the storage device takes place over SCSI or FC protocol. Compared to internal DAS, an external DAS overcomes the distance and device count limitations and provides centralized management of storage devices.

host

host

Storage Array

host

SAN:
A storage area network (SAN) carries data between servers (also known as hosts) and storage devices through fiber channel switches. A SAN enables storage consolidation and allows storage to be shared across multiple servers. It enables organizations to connect geographically dispersed servers and storage. A SAN provides the physical communication infrastructure and enables secure and robust communication between host and storage devices. The SAN management interface organizes connections and manages storage elements and hosts. In its earliest implementation, the SAN was a simple grouping of hosts and the associated storage that was connected to a network using a hub as a connectivity device. This configuration of a SAN is known as a Fiber Channel Arbitrated Loop (FC-AL), which is detailed later in the chapter. Use of hubs resulted in isolated FC-AL SAN islands because hubs provide limited connectivity and bandwidth. The inherent limitations associated with hubs gave way to high-performance FC switches. The switched fabric topologies improved connectivity and performance, which enabled SANs to be highly scalable. This enhanced data accessibility to applications across the enterprise. FC-AL has been abandoned for SANs due to its limitations, but still survives as a disk-drive interface. The FC SAN evolution from FC-AL to enterprise SANs. Today, Internet Protocol (IP) has become an option to interconnect geographically separated SANs. Two popular protocols that extend block-level access to applications over IP are iSCSI and Fiber Channel over.

se rv er

se rv er

se rv er

FC SAN

Storage Array

Storage Array

NAS:
Network-attached storage (NAS) is an IP-based file-sharing device attached to a local area network. NAS provides the advantages of server consolidation by eliminating the need for multiple file servers. It provides storage consolidation through file-level data access and sharing. NAS is a preferred storage solution that enables clients to share files quickly and directly with minimum storage management overhead. NAS also helps to eliminate bottlenecks that users face when accessing files from a general-purpose server. NAS uses network and file-sharing protocols to perform filing and storage functions. These protocols include TCP/IP for data transfer and CIFS and NFS for remote file service. NAS enables both UNIX and Microsoft Windows users to share the same data seamlessly. To enable data sharing, NAS typically uses NFS for UNIX, CIFS for Windows, and File Transfer Protocol (FTP) and other protocols for both environments. Recent advancements in networking technology have enabled NAS to scale up to enterprise requirements for improved performance and reliability in accessing data. A NAS device is a dedicated, high-performance, high-speed, single-purpose file serving and storage system. NAS serves a mix of clients and servers over an IP network. Most NAS devices support multiple interfaces and networks. A NAS device uses its own operating system and integrated hardware, software components to meet specific file service needs. Its operating system is optimized for file I/O and, therefore, performs file I/O better than a general purpose server. As a result, a NAS device can serve more clients than traditional file servers, providing the benefit of server consolidation.

Network attached storage (NAS)

N/w client

N/w client

N/w client

N/w client

NAS server
NAS private Storage

5. Benefits and limitations of DAS, NAS and CAS.


Benefits of DAS
Benefits of DAS include the following DAS requires a relatively lower initial investment than storage networking. DAS configuration is simple and can be deployed easily and rapidly. Setup is managed using host-based tools, such as the host OS, which makes storage management tasks easy for small and medium enterprises. DAS is the simplest solution when compared to other storage networking models and requires fewer management tasks, and less hardware and software elements to set up and operate.

Limitations of DAS
Limitation of DAS include the following DAS does not scale well. A storage device has a limited number of ports, which restricts the number of hosts that can directly connect to the storage. A limited bandwidth in DAS restricts the available I/O processing capability. When capacities are being reached, the service availability may be compromised and this has a ripple effect on the performance of all hosts attached to that specific device.

The distance limitations associated with implementing DAS because of direct connectivity requirements can be addressed by using Fibre Channel connectivity. DAS does not make optimal use of resources due to its limited ability to share front end ports. In DAS environments, unused resources cannot be easily re-allocated, resulting in islands of over-utilized and under-utilized storage pools. The absence of storage interconnects and network latency provide DAS with the potential to outperform other storage networking configurations.

Benefits of NAS
NAS offers the following benefits: Improved efficiency: Eliminates bottlenecks that occur during file access from a general-purpose file server because NAS uses an operating system specialized for file serving. It improves the utilization of general-purpose servers by relieving them of file-server operations. Improved flexibility: Compatible for clients on both UNIX and Windows platforms using industry-standard protocols. NAS is flexible and can serve requests from different types of clients from the same source. Centralized storage: Centralizes data storage to minimize data duplication on client workstations, simplify data management, and ensures greater data protection. Simplified management: Provides a centralized console that makes it possible to manage file systems efficiently. Scalability: Scales well in accordance with different utilization profiles and types of business applications because of the high performance and low-latency design. High availability: Offers efficient replication and recovery options, enabling high data availability. NAS uses redundant networking components that provide maximum connectivity options. A NAS device can use clustering technology for failover. Security: Ensures security, user authentication, and file locking in conjunction with industry-standard security schemas.

Limitation of NAS
Limitation of NAS include the following There is no guarantee that the security information associated with the files and directories will be processed. There will be performance degradation due to the fact that data is being accessed remotely. An NAS server, using FTP or SMBs, uses a simple, clear text exchange of account name/password. This isnt as secure as Kerberos because NAS devices which dont use NTFS may not support anything better than simple file sharing.

Certain NAS devices fail to expose well-known services that are typical of a file server, or enable them in a way that is not efficient. If the NAS is occupied with too many users, too many I/O operations, or CPU processing power that is too demanding, the NAS reaches its limitations. A server system is easily upgraded by adding one or more servers into a cluster, so CPU power can be upgraded, while the NAS is limited to its own hardware, which is in most cases not upgradeable.

Benefits of CAS
Benefits of CAS include the following: Content authenticity: It assures the genuineness of stored content. Content authenticity is assured because the address assigned to each piece of fixed content is as unique as a fingerprint. Content integrity: Refers to the assurance that the stored content has not been altered. Use of hashing algorithm ensures content integrity. Location independence: CAS uses a unique identifier that applications can leverage to retrieve data rather than a centralized directory, path names, or URLs. This yields complete content mobility to applications across locations. Single-instance storage (SiS): The unique signature is used to guarantee the storage of only a single instance of an object. This signature is derived from the binary representation of the object. Retention enforcement: Protecting and retaining data objects is a core requirement of an archive system. Record-level protection and disposition: All fixed content is stored in CAS once and is backed up with a protection scheme. Technology independence: The CAS system interface is impervious to technology changes. Fast record retrieval: CAS maintains all content on disks that provide sub second time to first byte (200 ms400 ms) in a single cluster.

Limitations of CAS
Limitations of CAS include the following: For data that changes frequently, CAS is not as efficient as location-based addressing. In these cases, the CAS device would need to continually re-compute the address of data as it was changed, and the client systems would be forced to continually update information regarding where a given document exists. For random access systems, a CAS would also need to handle the possibility of diverging two initially documents, requiring a copy of one document to be created on demand.

6. Study of industry management standards (SNMP, SMI-S, CIM).


1. SNMP
The most common system management infrastructure is based upon SNMP. The SNMP protocol was the standard used to manage Multi-vendor SAN environments. However, SNMP was primarily a network management protocol and was inadequate for providing the detailed information and functionality required to manage the SAN environment. SNMP definitions and data are represented in a text file called a management information Base (MIB). Modeling a new object in SNMP requires defining a new MIB. One MIB has no inherent relationship to any other MIB. A separate MIB is needed for each defined SNMP managed object. Achieving industry interoperability requires that each participating vendor agrees to use the same MIBs in the same way. One MIB has no inherent relationship to any other MIB. Although one MIB may contain some of the same properties as another MIB, no SNMP mechanism exists to take advantage of the commonality. Furthermore, SNMP cannot easily model objects in an enterprise or storage area network(SAN). SNMP was originally focused to manage only low level network components, not the entire enterprise.

Its naming scheme and infrastructure do not lend themselves to management of diverse enterprise environments. For example, SNMP cannot model non-computing elements such as corporate organizational units.

2. SMI-S
The storage Networking industry Association (SNIA) has been engaged in an initiative to develop a common, open storage, and SAN management interface. SMI-S is based on Web-Based enterprise management technology and the DMTFs common information model (CIM). The initiative was formally created to enable broad interoperability among heterogeneous storage vendor systems. And to enable betted management solutions that span these environments. This initiative is known as the storage management initiative. The storage management initiative chose the common information model because CIM is a much more extensible data model.

The SMI chose web based enterprise management as the underlying infrastructure technology because it supplied a richer set of capabilities. SMIS-s will support a number of essential SAN management functions in an open, device-neutral way: discovering, monitoring, and device management. The SMI specification known as SMI-s. SMI is a model, object-oriented, command and control oriented and its provides a single unified view of a SAN.

3. CIM
The common information model is an open standard that defines how managed elements in an IT environment are represented as a common set of objects and relationships between them. This is intended to allow consistent management of these managed elements, independent of their manufacturer or provider. CIM allows multiple parties to exchange management information about these managed elements. However, CIM not only represents these managed elements and the management information, but also provides means to actively control and manage these elements. By using a common model of information management software can be written once and work with many implementations of the common model without complex and costly conversion operations or loss of information.

The CIM standard is defined and published by the distributed management Task force. A related standard is web based enterprise management , which defines a particular implementation of CIM, including protocols for discovering and accessing such CIM implementations.

7. Study of cloud computing Evolution.


The evolution of cloud computing is shown below:

Cloud Computing is a wonderful buzzword being touted by any and all in the technology industry these days although the definition varies widely depending on the source or sector. So what is Cloud Computing? At the recent Irish Open Source Technology Conference, we delivered a presentation (embedded below) called Cloud Computing The Evolution of Internet Infrastructure, which aimed to explain Cloud Computing, talk

about the components of a Cloud Platform, and then discuss the future potential of this area. Introduction to Cloud Computing Cloud Computing is often confused with Grid Computing or Utility Computing, and while these may be uses of Cloud Computing, they are not inherently part of it. Grid Computing is typically a massively scaled infrastructure capable of processing huge amounts of data very quickly. Environmental Modeling or DNA Modeling are examples of uses for grid computing. Utility computing on the other hand is really a model of pricing and delivery. In utility computing, the premise is that you can access computing resources as you need them, and only pay as you use them. Utility Computing is often explained by referencing the Electricity Grid where there is a massive power grid that you can tap into as you need it, and scale up and down as you require.Cloud Computing is essentially computing resources delivered from the Internet to your desktop. Cloud Computing refers to services on the Internet as being in the cloud which means not in your office server rack or your buildings comms room. Cloud Computing means outsourcing your computing requirements and receiving these services through your Internet connection. You no longer need to worry about power, cooling, server hardware deployment and management.You can use Cloud Computing services to store data (archive, backup, general-purpose), deliver applications (SaaS Software-as-a-Service), and deliver access to Internet Infrastructure as and when your company requires it (HaaS Hardware-as-aService), such as for seasonal peaks, or torture-testing your new application deployment. Why use Cloud Computing? Let me preface this by assuming there is an organization need for Internet Infrastructure be that a requirement to delivery services to customers (IE a web application your company has built), or to provide resources to staff such as email, file storage, collaboration, database and ERP applications. Assuming an Infrastructure requirement exists, there are two ways it may be currently at the moment. You may have infrastructure in-house (typically for office applications) or you may have server hardware co-located or leased in a Data Centre (typically for web applications). Enter Cloud Computing. First of all, the most tangible benefit of Cloud Computing is that it offers an Enterprise-grade solution, on a very affordable basis. Unlike a single server, an enterprise cloud offering will have massive amounts of fault tolerance built in, and should be architected using high-end network, server and storage kit. You can therefore outsource your companys infrastructure to a Cloud Platform and benefit from reduced cost AND increased services. (This is why Cloud Computing is such a hyped area!). Some other rationale for choosing Cloud Computing over traditional deployments includes:

Scalability: Grow your deployment rapidly, as required, without huge capital costs or operational time. Flexibility: Add and remove resources on-the-fly (or script tolerances!) to cope with peaks in your requirements. Only pay for what you need. Reliability: Take advantage of a massive computing platform, without having to build and buy your own. Improve your organizations infrastructure SLAs by using a highly redundant and resilient platform that has no single points of failure (i.e. that server in the corner of your office!) Fast Setup: React the decisions or requirements quickly and deploy complex architectures rapidly. Affordable Enterprise Solution: Economies of Scale. Utilize a multi-million euro platform for even a single server deployment to get best-in-class service for an entry level price. Environmentally Efficient: Cloud Platforms maximize infrastructure utilization per server and per square foot in the Data Centre. Standard servers only use 15% of their resources.

The 3 Levels of Cloud Computing 1. Hardware Independence 2. Service Not Reliant on single Data Centre

8. Define cloud vocabulary.


Cloudburst The term cloudburst is being use in two meanings, negative and positive: 1. Cloudburst (negative): The failure of a cloud computing environment due to the inability to handle a spike in demand. 2. Cloudburst (positive): The dynamic deployment of a software application that runs on internal organizational compute resources to a public cloud to address a spike in demand. 3. Cloudstorming: The act of connecting multiple cloud computing environments. 4. Vertical Cloud: A cloud computing environment optimized for use in a particular vertical -- i.e., industry -- or application use case. 5. Private Cloud: A cloud computing-like environment within the boundaries of an organization and typically for its exclusive usage. 6. Internal Cloud: A cloud computing-like environment within the boundaries of an organization and typically available for exclusive use by said organization. 7. Hybrid Cloud: A computing environment combining both private (internal) and public (external) cloud computing environments. May either be on a continuous basis or in the form of a 'cloudburst'. 8. Cloudware: A general term referring to a variety of software, typically at the infrastructure level, that enables building, deploying, running or managing applications in a cloud computing environment.

9. External Cloud: A cloud computing environment that is external to the boundaries of the organization. Although it often is, an external cloud is not necessarily a public cloud. Some external clouds make their cloud infrastructure available to specific other organizations and not to the public at-large. 10. Public Cloud: A cloud computing environment that is open for use to the general public, whether individuals, corporations or other types of organizations. Amazon Web Services are an example of a public cloud. 11. Cloud Provider: An organization that makes a cloud computing environment available to others, such as an external or public cloud. 12. Cloud Enabler: A general term that refers to organizations (typically vendors) who are not cloud providers per se, but make available technology, such as cloudware, that enables cloud computing. 13. Cloud-Oriented Architecture (COA): An architecture for IT infrastructure and software applications that is optimized for use in cloud computing environments. The term is not yet in wide use, and as is the case for the term "cloud computing" itself, there is no common or generally accepted definition or specific description of a cloud-oriented architecture. 14. Cloud Service Architecture (CSA): A term coined by Jeff Barr, chief evangelist at Amazon Web Services. The term describes an architecture in which applications and application components act as services on the cloud, which serve other applications within the same cloud environment.

15. Virtual Private Cloud (VPC): A term coined by Reuven Cohen, CEO and founder of Anomaly. The term describes a concept that is similar to, and derived from, the familiar concept of a Virtual Private Network (VPN), but applied to cloud computing. It is the notion of turning a public cloud into a virtual private cloud, particularly in terms of security and the ability to create a VPC across components that are both within the cloud and external to it. 16. Cloud Portability: The ability to move applications (and often their associated data) across cloud computing environments from different cloud providers, as well as across private or internal cloud and public or external clouds. 17. Cloud sourcing - As defined by Dion Hinchcliffe: "Leveraging services in the network cloud to provide external computing capabilities, often to replace more expensive local IT capabilities. Cloud sourcing can theoretically provide significant economic benefits along with some attendant trade-offs. These trade-offs can include security and performance. The term "cloud" represents a set of external services on a 3rd party network, usually the Internet. The services can represent raw computing, storage, messaging, or more structured capabilities such as vertical and horizontal business applications, even community. These services are delivered over the network, but generally behave as if they are local." 18. Cloud Spanning: Running an application in a way that its components straddle multiple cloud environments (which could be any combination of internal/private and external/public clouds. Unlike Cloud Bursting, which refers strictly to expanding the application to an External Cloud to handle spikes in demand, Cloud

Spanning includes scenarios in which an applications component are continuously distributed across multiple clouds.

9. Study of Amazon cloud.


Amazon EC2 Cloud Computing Service: Amazon EC2 is a popular elastic cloud
computing service that has developers in mind. EC2 offers complete control over web computing resources and structures its pay-as-you-go feature as specific as a per minute usage basis. Amazon offers other web services that allow developers to easily utilize different degrees of infrastructure and SaaS needs on an as-needed basis. Use EC2 with Other Amazon Web Services. Amazon EC2 works with other Amazon services such as Amazon Simple Storage Service (Amazon S3), Amazon Simple DB and Amazon Simple Queue Service (Amazon SQS). The Amazon Machine Image (AMI) allows users to utilize applications, create libraries, data, and associated configuration settings. It is also possible to use template images that are already available for usage and allow instance usage of EC2. In comparison to other cloud computing services such as Microsoft Azure, Go Grid, and Google App Engine, Amazon EC2 offers a very flexible structure that works with almost every platform. Amazon EC2 has also been around for a few years already and so has had more of an opportunity to claim market share and presence in the Cloud Computing community.

Amazon EC2 Cloud Computing Case Study: Amazon has quite a few case
studies available that go in-depth regarding the actual monetary savings and results from

businesses using their web services. Although there are additional case studies available on the Amazon site, here is an example of a case study that had a lot of impact on the organization financially: Mailtrust Company Profile: 100% focused on business email hosting and at the time of the case study, they were managing more than 800,000 paid business email accounts. Mailtrust has since been acquired by Rack space and is now the Email and Apps division of Rack space Hosting, which handles more than 1 million accounts. What did EC2 do for them?: Amazon EC2 saved them over $187,000 and MailTrust projects their savings for year two at $545,000 and savings for year three at $1.3 Million. The company projected that this would save the company around $2 Million in three years. Mailtrust used Amazon Simple Storage Service (Amazon S3) for a faster, more costefficient online storage solution in conjunction with Amazon Elastic Compute Cloud (Amazon EC2) and Amazon Simple Queue Service (Amazon SQS).

Amazon S3 Storage has helped us lower costs, improve performance of our email hosting system, and enhance the reliability of our email data backups, and has allowed us to deploy resources in more strategic ways, said Patrick Matthews, president of Mailtrust. With Amazon S3, Mailtrust now has a fully scalable, high performance backup system. They customized their backup client to eliminate duplication, they backup all servers in parallel, and they run backups at night to avoid slow email performance. And with the power of Amazons API, they built a web-based tool that enabled customers to browse their backups and perform their own data restores for free.

10. Future scope of cloud computing.


With arrival of cloud computing the conventional way of computing has gone for a sea change. As per some expert opinions, it is going to be the face of future cloud computing. And hence, the future of cloud computing seems very promising. Let us have a look on what is makes the future of cloud computing so bright.

The Future of Cloud Computing


With arrival of cloud computing the conventional way of computing has gone for a sea change. And this new addition in the computing is not a flash in the pan as it is going to rule the roost in the future. As per some expert opinions, it is going to be the face of future cloud computing. And hence, the future of cloud computing seems very promising. As per some surveys conducted by leading organizations, 70% of Americans will be getting benefited from cloud and it various applications in next decades for official and personal use. And this is not an overestimate or exaggeration, as we are already using cloud and its applications in one form or another. Using email and connecting to social media through smart phones, watching movies over smart phones and uploading and accessing

pictures from websites like Flicker are common examples of cloud computing in our dayto-day life. Let us have a look on what makes the future of cloud computing so bright. 1. Presence of Internet will boost its future: The cloud computing will becomes all the more important with the omnipresence of high-speed, broadband Internet. Slowly but steadily we are getting closer. Even airlines are offering satellite based wi-fi services in flights. In a mass drive to connect every village with Internet wireless Internet services are offered through the help of satellite, although speed is a bit slow. This increasing presence of Internet is opening new vistas in educations and healthcare. With the help of cloud computing we can use these services at a little cost. 2. No more software updates: Most of the computer professionals loose lots of their time and efforts downloading different versions of software so that they can access the various programs and data with little efforts. Most of the softwares are on the cloud servers so you don't need to down and install for little use. So, whether you want to access emails or go through spreadsheet, it has become fun with the arrival of cloud computing. As per some estimates a sizable number of softwares will be delivered through the Internet. 3. Hardware optional: With the arrival of cloud computing it is no longer necessary to purchase hard drives with large storage capacity, as it can be stored on cloud. So keep the fear of loosing your data away. All your data with complete back up can be stored on the cloud. So with rising popularity the computers will act as an interface to communicate with cloud Computing.

4. Entertainment unlimited: As hardware is no more mandatory, so there is no limit on entertainment options. Uploading latest software and buying games from the market is going to be things of the past. In the future, there will be mobile 3D games to entertain your kids. 5. Medical treatments simplified: The future of cloud computing is not confined to entertainment and gaming options as it can contribute massively in the fields of medical sciences as well. As most of the contemporary treatment require computer assistance, as data have to be searched for various things like DNA samples and other biochemical procedures and hence cloud computing is going to play its part in the most of the therapies. In addition, it will make easy the task of data processing. 6. Weather Forecasting: It is believed that with increased level of computing coupled with improved climate models it will be lot more easy makes weather forecasts. 7. Education for all: With lot of educational institutions offering free course material for everyone over the Internet it is here cloud computing can play a great role delivering education on the doorsteps of learners over an interface. In addition, it will be a giant leap towards digitalization of education. So what if you have not

secured admission at a reputed university, you can learn various things over computer with Internet connection. 8. Freedom from Wallets: With the advent of mobile phones the concept of traditional wallets has gone for toss. Now everything right from your contact details to your shopping related needs and you air ticket for vacations to clicking the pictures of happy moments, everything can be done by you smart phone. The cloud has made it possible. It is possible in the futures that you can store all you valuable documents like driving license and voter identity with the help of your smart phone. 9. Paperless society? Yes, it is possible.: You have seen pile of wastepaper in the offices or dustbins full of torn paper. Now these things are going to change. There is no need to cut trees for making paper as we can do most of the transactions and communications online. And cloud computing has certainly a role to play here. Now we can book mobile tickets for concerts and moving, as most of these things are available on the cloud. 10. No need to rub shoulders: If going to a grocery store irritates you, as it is crowded don't go there. Just get to the Internet and add things to the cart and order. The product will be delivered at your doorstep. Nowadays the shopping has gone online and cloud computing has certain role to play in the business. Are you still doubtful about the future of cloud computing? 11. Get your location: The location services offered by some social networking sites like twitter in US and Foursquare assist helps people to locate their family and friends. With the help of cloud computing, the locations services are going to be better for sure. Now it can be used in rescuing operations to finding the location of the victims.

12. A boon to Digital media: Arrival of cloud could be a boon for the digital media. Now independent artists and creative writer can reach to more and more peoples and thus ending the monopoly of certain media organizations. The rising competition will open the floodgates of the creative writers and digital media content providers. If all the content is not for free the users can at least find out from where to by this music or a book and thus saving their lot of time and efforts. In the coming years, buying DVD from a market and endless cue at a cinema hall is going to be obsolete. 13. Cloud Computing: The New Age lifesaver: It may appears ridiculous and exaggerated to many but with the arrival of computing one can store the medical records on cloud in a single digital repository and one can access it from the any part of the world. And you don't have to lament that you have left you all medical records behind. Do you think it is possible to ECG, X-rays and blood test report while you are heading for a vacation? Enjoy your life on the vacation and in the case of some medical emergency you can access it by pressing some keys on keyboard and clicking the mouse. 14. A new age safety system: With the help of cloud computing, records of cars including number. Driving license and address detail of the owner can be stored on the cloud in case the car is stolen one and recovered by some security agency in far

off locations, they can instantly inform the owner of the car. It also lessens the burden of the various police and security organizations in a given region. Like wise complete details of all the population of a county that including Fingerprint and DNA records can be given. It will help recruiters in doing background checks before giving employment to candidates. It will help reduces theft and misconduct in some cases. 15. An efficient tool in disaster management: With improved location services the disaster management team find it easy to help the victims in the time of emergency. As we all know the golden hour is wasted in finding the locations of the victims. This facility will prove a boon in helping those in the need. Most of the times, when an aircraft meets an accident, it is hard to find the location of the wreckage of the aircraft. The cloud computing in conjunction to Satellite and aerial imagery can help find the location in a very short span of time and thus saving lot of time and effort of manual search done by security agencies. 16. Now everyone is entrepreneur: With arrival of cloud anyone can turn entrepreneur. It provides easy solution to various IT problems faced by of new startups. For example if you have a particular skill on display you can give it on cloud and likewise the artist can give samples of their art on You Tube, where they can find numerous potential takers for their skill. In addition, the role of cloud computing is set to increase in areas like warfare and new age gizmos and gadget in addition to various home appliances. Most of us often get trapped in traffic congestion on the road with improved technology one can find a path to avoid traffic snarls. Are you still cynic about the future of cloud computing?

INDEX

1.

Evolution of various storage technologies. Study of Information Lifecycle Management. Study of Disk physical structure components, properties and performance. Study of DAS, NAS, SAN. Security and limitations of DAS, NAS, CAS. Study of Industry management standards (SNMP, SMI-S, CIM). Study of Cloud computing Evolution. Cloud Vocabulary. Study cloud. of amozone

2.

3.

4.

5.

6.

7. 8. 9. 10.

Future scope of cloud.

INDEX

1. Draw Object, state, Data flow Diagram of ATM. 2.Draw event trace, event flow and state diagram of Telephone Call. 3.Draw event flow and event trace diagram of library information system. 4. Draw Object, state, context level Data flow Diagram of Library Information System. 5. Draw Object, state, Data flow Diagram of Online Sales.

You might also like