Professional Documents
Culture Documents
GRID
4. Introduction to Grid and its evolution
5. Implementing Production Grids
6. Anatomy of Grid
CLOUD
7. Introduction to Cloud Computing
8. Nature of Cloud
9. Cloud Elements
CLUSTER COMPUTING
Cluster Computing at Glance :
• Eras of Computing
• Scalable Parallel Computer Architecture
• Towards Low Cost Parallel Computing & Motivation
• Windows opportunity
• A Cluster Computer And Its Architecture
• Cluster Classification
• Commodity Components for Clusters
• Network Services/Communication SW
• Cluster Middleware and
• Single Systems Image
• Resource management & Scheduling (RMS),
• Programming environment Tools
• Cluster Applications
• Representative Cluster Systems
• Clusters of SMPS
[1] G. Pfister. In Search of Clusters. Prentice Hall PTR, NJ, 2nd Edition, NJ, 1998.
CLUSTER COMPUTING : ERAS OF COMPUTING
The technological
advances in hardware
(Chip development and
fabrication
technologies, fast and
cheap microprocessors,
high bandwidth and low
latency interconnection
networks)
The technological
advances in software
(Mature software, such
as Operating Systems,
programming
languages, development
methodologies, and
tools)
• Massively Parallel
Processors (MPP)
• Symmetric
Multiprocessors
(SMP)
• Cache-Coherent
Non-uniform
Memory Access
(CC-NUMA)
• Distributed
Systems
• Clusters
CLUSTER COMPUTING
DEFINITION
Individual workstations are becoming increasingly powerful. That is, workstation performance
has increased dramatically in the last few years and is doubling every 18 to 24 months. This is
likely to continue for several years, with faster processors and more efficient multiprocessor
machines coming into the market.
The communications bandwidth between workstations is increasing and latency is decreasing as
new networking technologies and protocols are implemented in a LAN.
Workstation clusters are easier to integrate into existing networks than special parallel
computers.
Typical low user utilization of personal workstations.
The development tools for workstations are more mature compared to the contrasting proprietary
solutions for parallel computers, mainly due to the nonstandard nature of many parallel systems.
Workstation clusters are a cheap and readily available alternative to specialized high
performance computing platforms.
Clusters can be easily grown; node's capability can be easily increased by adding memory or
additional processors.
CLUSTER COMPUTING
ADVANTAGES OF CLUSTER OVER SPECIALIZED COMPUTERS
Clusters offer the following benefits at a relatively low cost:
High Performance
Expandability and Scalability
High Throughput
High Availability
Failover capability
CLUSTER COMPUTING
RESEARCH OPPORTUNITIES
Parallel Processing - Use the multiple processors to build MPP/DSM-like
systems for parallel computing.
Individual clusters may be interconnected to form a larger system (clusters of clusters) and, in
fact, the Internet itself can be used as a computing cluster. The use of wide-area networks of
computer resources for high performance computing has led to the emergence of a new field
called Meta-computing.
CLUSTER COMPUTING
COMMODITY COMPONENTS FOR CLUSTERS
Intel processors are most commonly used in PC-based computers. The current
generation Intel x86 processor family matches the performance of medium level
workstation processors. (In the high performance range, the Pentium Pro shows a very
strong integer performance, beating Sun's Ultra SPARC at the same clock speed;
however, the floating-point performance is much lower.) The Pentium II Xeon, uses a
100 MHz memory bus and available with a choice of 512KB to 2MB of L2 cache. The
accompanying 450NX chipset for the Xeon supports 64-bit PCI busses that can support
Gigabit interconnects. Other popular processors include x86 variants (AMD x86,
Cyrix x86), Digital Alpha, IBM PowerPC, Sun SPARC, SGI MIPS, and HP PA.
Computer systems based on these processors have also been used as clusters; for
example, Berkeley NOW uses Sun's SPARC family of processors in their cluster
nodes.
CLUSTER COMPUTING
COMMODITY COMPONENTS FOR CLUSTERS
MEMORY AND CACHE
Originally, the memory present within a PC was 640 KB, usually 'hardwired' onto the
motherboard. Typically, a PC today is delivered with between 32 and 64 MB installed in slots
with each slot holding a Standard Industry Memory Module (SIMM). Computer systems can use
various types of memory and they include Extended Data Out (EDO) and fast page. EDO allows
the next access to begin while the previous data is still being read, and fast page allows multiple
adjacent accesses to be made more efficiently.
The amount of memory needed for the cluster is likely to be determined by the cluster target
applications. Programs that are parallelized should be distributed such that the memory, as well
as the processing, is distributed between processors for scalability. Thus, it is not necessary to
have a RAM that can hold the entire problem in memory on each system, but it should be enough
to avoid the occurrence of too much swapping of memory blocks (page-misses) to disk, since
disk access has a large impact on performance.
Access to DRAM is extremely slow compared to the speed of the processor. Caches are used to
keep recently used blocks of memory for very fast access if the CPU references a word from that
block again. However, the very fast memory used for cache is expensive and cache control
circuitry becomes more complex as the size of the cache grows. Because of these limitations, the
total size of a cache is usually in the range of 8KB to 2MB.
CLUSTER COMPUTING
COMMODITY COMPONENTS FOR CLUSTERS
DISK AND I/O
Now, the state-of-the-art Ethernet is the Gigabit Ethernet and its attraction is largely
due to two key characteristics. First, it preserves Ethernet's simplicity while enabling a
smooth migration to Gigabit-per-second (Gbps) speeds. Second, it delivers a very high
bandwidth to aggregate multiple Fast Ethernet segments and to support high-speed
server connections, switched intrabuilding backbones, inter-switch links, and high-
speed workgroup networks.
CLUSTER COMPUTING
COMMODITY COMPONENTS FOR CLUSTERS
CLUSTER INTERCONNECTS
Asynchronous Transfer Mode (ATM)
ATM is a switched virtual-circuit technology and was originally developed for the
telecommunications industry. It is embodied within a set of protocols and standards.
Unlike some other networking technologies, ATM is intended to be used for both LAN
and WAN, presenting a unified approach to both. ATM is based around small fixed-size
data packets termed cells. It is designed to allow cells to be transferred using a number
of different media such as both copper wire and fiber optic cables. However, this is
undesirable in desktop environments; for example, twisted pair cables may have been
used to interconnect a networked environment and moving to fiber-based ATM would
mean an expensive upgrade. The two most common cabling technologies found in a
desktop environment are telephone style cables (CAT-3) and a better quality cable
(CAT-5). CAT-5 can be used with ATM allowing upgrades of existing networks without
replacing cabling.
CLUSTER COMPUTING
COMMODITY COMPONENTS FOR CLUSTERS
CLUSTER INTERCONNECTS
Scalable Coherent Interface (SCI)
SCI aimed at providing a low-latency distributed shared memory across a cluster. SCI
is the modern equivalent of a Processor-Memory-I/O bus and LAN combined. It is
designed to support distributed multiprocessing with high bandwidth and low latency.
It provides a scalable architecture that allows large systems to be built out of many
inexpensive mass-produced components.
Although SCI is favored in terms of fast distributed shared memory support, it has not
been taken up widely because its scalability is constrained by the current generation of
switches and its components are relatively expensive.
CLUSTER COMPUTING
COMMODITY COMPONENTS FOR CLUSTERS
CLUSTER INTERCONNECTS
Myrinet
Myrinet is a 1.28 Gbps full duplex interconnection network supplied by Myricom. Myrinet uses
low latency cut-through routing switches, which is able to offer fault tolerance by automatic
mapping of the network configuration. Myrinet supports both Linux and NT. In addition to
TCP/IP support, the MPICH implementation of MPI is also available on a number of custom-
developed packages such as Berkeley active messages, which provide sub-10 µs latencies.
Myrinet is relatively expensive when compared to Fast Ethernet, but has real advantages over it:
very low-latency (5 µs, one-way point-to-point), very high throughput, and a programmable on-
board processor allowing for greater flexibility. It can saturate the effective bandwidth of a PCI
bus at almost 120 Mbytes/s with 4Kbytes packets.
One of the main disadvantages of Myrinet is, as mentioned, its price compared to Fast Ethernet.
The cost of Myrinet-LAN components, including the cables and switches, is in the range of
$1,500 per host. Also, switches with more than 16 ports are unavailable, so scaling can be
complicated, although switch chaining is used to construct larger Myrinet clusters.
CLUSTER COMPUTING
NETWORK SERVICES/COMMUNICATION S/W
The communication needs of distributed applications are diverse and varied and range from reliable point-to-
point to unreliable multicast communications. The communications infrastructure needs to support protocols
that are used for bulk-data transport, streaming data, group communications, and those used by distributed
objects.
The communication services employed provide the basic mechanisms needed by a cluster to transport
administrative and user data. These services will also provide the cluster with important quality of service
parameters, such as latency, bandwidth, reliability, fault-tolerance etc. Typically, the network services are
designed as a hierarchical stack of protocols.
Traditionally, the operating system services (pipes/sockets) have been used for communication between
processes in message passing systems. As a result, communication between source and destination involves
expensive operations, such as the passing of messages between many layers, data copying, protection
checking, and reliable communication measures. Often, clusters with a special network/switch like Myrinet
use lightweight communication protocols such as active messages for fast communication among its nodes.
They potentially bypass the operating system and thus remove the critical communication overheads and
provide direct, user-level access to the network interface.
Often in clusters, the network services will be built from a relatively low-level communication API
(Application Programming Interface) that can be used to support a wide range of high-level communication
libraries and protocols.
CLUSTER COMPUTING
CLUSTER MIDDLEWARE AND SINGLE SYSTEM IMAGE
It frees the end user from having to know where an application will run.
It frees the operator from having to know where a resource (an instance of resource) is
located.
It does not restrict the operator or system programmer who needs to work on a
particular region; the end user interface (hyperlink - makes it easy to inspect
consolidated data in more detail) can navigate to the region where a problem has arisen.
It reduces the risk of operator errors, with the result that end users see improved
reliability and higher availability of the system.
CLUSTER COMPUTING
CLUSTER MIDDLEWARE AND SINGLE SYSTEM IMAGE
It allows to centralize/decentralize system management and control to avoid the need
of skilled administrators for system administration.
It helps track the locations of all resources so that there is no longer any need for
system operators to be concerned with their physical location while carrying out system
management tasks.
The benefits of a SSI also apply to system programmers. It reduces the time, effort and
knowledge required to perform tasks, and allows current staff to handle larger or more
complex systems.
CLUSTER COMPUTING
CLUSTER MIDDLEWARE AND SINGLE SYSTEM IMAGE
SINGLE SYSTEM IMAGE LEVELS/LAYERS
The SSI concept can be applied to applications, specific subsystems, or the
entire server cluster. Single system image and system availability services can
be offered by one or more of the following levels/layers:
A full cluster-wide SSI allows all physical resources and kernel resources to be visible
and accessible from all nodes within the system. Full SSI can be achieved as underware
(SSI at OS level). In other words, each node's OS kernel cooperating to present the
same view from all kernel interfaces on all nodes.
The full SSI at kernel level, can save time and money because existing programs and
applications do not have to be rewritten to work in this new environment. In addition,
these applications will run on any node without administrative setup, and processes can
be migrated to load balance between the nodes and also to support fault-tolerance if
necessary.
CLUSTER COMPUTING
CLUSTER MIDDLEWARE AND SINGLE SYSTEM IMAGE
SINGLE SYSTEM IMAGE LEVELS/LAYERS
APPLICATIONS AND SUBSYSTEMS LAYER (MIDDLEWARE)
SSI can also be supported by applications and subsystems, which presents multiple,
cooperating components of an application to the user/administrator as a single
application. The application level SSI is the highest and in a sense most important,
because this is what the end user sees. For instance, a cluster administration tool offers
a single point of management and control SSI services. These can be built as GUI-
based tools offering a single window for the monitoring and control of cluster as a
whole, individual nodes, or specific system components.
The subsystems offer a software means for creating an easy-to-use and efficient cluster
system. Run time systems, such as cluster file systems, make disks attached to cluster
nodes appear as a single large storage system. SSI offered by file systems ensures that
every node in the cluster has the same view of the data. Global job scheduling systems
manage resources, and enables the scheduling of system activities and execution of
applications while offering high availability services transparently.
CLUSTER COMPUTING
CLUSTER MIDDLEWARE AND SINGLE SYSTEM IMAGE
SSI BOUNDARIES
A key that provides structure to the SSI lies in noting the following points:
Every single system image has a boundary; and
Single system image support can exist at different levels within a system-one able to
be built on another.
For instance, a subsystem (resource management systems like LSF and CODINE) can
make a collection of interconnected machines appear as one big machine. When any
operation is performed within the SSI boundary of the subsystem, it provides an
illusion of a classical supercomputer. But if anything is performed outside its SSI
boundary, the cluster appears to be just a bunch of connected computers. Another
subsystem/application can make the same set of machines appear as a large
database/storage system. For instance, a cluster file system built using local disks
associated with nodes can appear as a large storage system (software RAID)/parallel
file system and offer faster access to the data.
CLUSTER COMPUTING
CLUSTER MIDDLEWARE AND SINGLE SYSTEM IMAGE
MIDDLEWARE DESIGN GOALS
The design goals of cluster-based systems are mainly focused on complete
transparency in resource management, scalable performance, and system availability in
supporting user applications.
Complete Transparency: The SSI layer must allow the user to use a cluster easily and
effectively without the knowledge of the underlying system architecture.
Enhanced Availability: The middleware services must be highly available at all times.
At any time, a point of failure should be recoverable without affecting a user's
application. This can be achieved by employing check pointing and fault tolerant
technologies (hot standby, mirroring, failover, and failback services) to enable rollback
recovery.
CLUSTER COMPUTING
CLUSTER MIDDLEWARE AND SINGLE SYSTEM IMAGE
KEY SERVICES OF SSI AND AVAILABILITY INFRASTRUCTURE
• Threads
• Message Passing Systems (MPI and PVM)
• Distributed Shared Memory (DSM) Systems
• Parallel Debuggers and Profilers
• Performance Analysis Tools
• Cluster Administration Tools
CLUSTER COMPUTING
PROGRAMMING ENVIRONMENTS AND TOOLS
THREADS
This has two advantages: It is possible to benefit from the high performance, easy-to-
use-and-program SMP systems with a small number of CPUs. In addition, clusters can
be set up with moderate effort (for example, a 32-CPU cluster can be constructed by
using either commonly available eight 4-CPU SMPs or four 8-CPU SMPs instead of 32
single CPU machines) resulting in easier administration and better support for data
locality inside a node.
This trend puts a new demand on cluster interconnects. For example, a single NIC will
not be sucient for an 8-CPU system and will necessitate the need for multiple network
devices. In addition, software layers need to implement multiple mechanisms for data
transfer (via shared memory inside an SMP node and the network to other nodes).
CLOUD COMPUTING
1. Introduction to Cloud Computing
• Defining clouds
• Cloud Providers
• Consuming Cloud Services
• Cloud Models (IaaS, PaaS, SaaS)
• Inside the cloud
• Administering Cloud Services
• Technical Interface
• Cloud Resources
2. Nature of Cloud
• Traditional data center
• Cost of cloud data center
• Scaling computer systems
• Economics
• Cloud Work Load
• Managing data on cloud
• Public, Private and Hybrid Cloud
3. Cloud Elements
• IaaS
• PaaS
• SaaS
INTRODUCTION TO CLOUD COMPUTING
• Defining clouds
• Cloud Providers
• Consuming Cloud Services
• Cloud Models (IaaS, PaaS, SaaS)
• Inside the cloud
• Administering Cloud Services
• Technical Interface
• Cloud Resources
EVOLUTION OF COMPUTING
Mainframe
↓
PC
↓
Client/Server
↓
Web Services
↓
Cloud
CLOUD COMPUTING
HOW EXPERTS SEE IT!
In India, cloud computing is projected to grow from a $50 million industry in 2009 to a $15
billion industry by 2013.[1]
“It has the potential to create a paradigm shift in the way IT resources are used and distributed,” -
P.K.Sinha, chief coordinator for research and development at the Center for Development of
Advanced Computing at Pune University in India.
“An organization’s IT infrastructure may become inaccessible for some time or data sitting in the
cloud could be lost,” - Gautam Shroff, vice president at Tata Consultancy Services Ltd. in Delhi,
India
IBM has established cloud computing centers in Beijing, China; Bangalore, India; Hanoi,
Vietnam; Sao Paulo, Brazil; and Seoul, South Korea.
In India, the Apparel Export Promotion Council has developed a cloud platform that provides
computing services and enterprise software to more than 11,000 of its members— most of whom
lack the capital and resources to build their own IT infrastructure.
[1] Samuel Greengard, Cloud Computing and Developing Nations, © 2010 ACM 0001-0782/10/0500
CLOUD COMPUTING
HOW EXPERTS SEE IT!
Cloud computing is not so much a definition of a single term as a trend in service
delivery. It’s the movement of application services onto the Internet and the increased
use of the Internet to access a variety of services traditionally originating from within a
company’s data center.- Lew Tucker[1]
If you run your services inside the company, privately, utilization becomes an issue. It
amortizes your costs over a number of cycles. If you run services outside, on a public
service, it is no longer an issue for you. - Werner Vogels[1]
[1] CTO Roundtable: Cloud Computing, communications of the acm, august 2009, vol. 52, no. 8
INTRODUCTION TO CLOUD COMPUTING
Defining clouds:
• The cloud itself is a set of hardware, networks, storage, services, and
interfaces that enable the delivery of computing as a service. Cloud
services include delivery of software, infrastructure, and storage over
Internet based on user demand. (Source: Text Book)
• Cloud computing refers to the use and access of multiple server-based
computational resources via a digital network (WAN, Internet connection
using the World Wide Web, etc.). Cloud users may access the server
resources using a computer, notebook, pad computer, smart phone, or
other device. In cloud computing, applications are provided and
managed by the cloud server and data is also stored remotely in the cloud
configuration. Users do not download and install applications on their
own device or computer; all processing and storage is maintained by the
cloud server. The on-line services may be offered from a cloud provider
or by a private organization. (Source: http://en.wikipedia.org/wiki/Cloud_computing)
INTRODUCTION TO CLOUD COMPUTING
Defining clouds:
– An Internet based computing environment where you pay only for
resources that you use
Defining clouds:
In SaaS, a complete application is offered as service on demand.
The service is hosted on Internet so that it runs on a server and
accessed through web interface. (e.g. salesforce.com, Google
Apps).
Defining clouds:
Clouds intend to control the next generation data centers by
building them as a network of virtual services so that users can
access and deploy applications from anywhere on demand at viable
expenses depending on consumers' Quality of Service requirements
or Service Level Agreements (SLA). Developers are no longer
required to make large capital outlays in the hardware and software
infrastructures to deploy their services or human expense to operate
it. It offers significant benefit to IT companies by freeing them
from the low level task of setting up basic hardware and software
infrastructures and thus enabling more focus on core business.
INTRODUCTION TO CLOUD COMPUTING
Cloud Features:
– The National Institute of Standards and Technology says there are
five key features of cloud computing: on-demand self-service,
ubiquitous network access, location-independent resource pooling,
rapid elasticity, and measured service (Wyld, 2009, p. 12)
– As per text book, there are four features of Cloud computing.
o Elastic and ability to scale up/down
o Self-service provisioning and automatic de-provisioning
o Application programming interfaces (APIs)
o Billing and metering of service usage in a pay-as-you-go model.
INTRODUCTION TO CLOUD COMPUTING
Cloud Features
Elasticity: Customers’ usage may vary over a period of time. Service should be
made available 24 x 7, and should scale upward for busy period and downward
for slag period.
Scalability: In case of additional users or change in application requirement,
application should scale up/down.
Self-Service Provisioning: Customer should be able to get what he expects
without putting himself into too much technical details.
API: Interfaces in form of standardized API provide the guideline on how two
applications or data sources can communicate with each other.
Billing & Metering Services: Usage of computing resources needs to be metered
(tracked).
Performance Monitoring & Measuring: CSP must provide service management
environment which manages resources involved. Key issues to be considered by
this environment are: (a) service monitoring (b) security (c) performance
(uptime/downtime)
Security: As user data is stored on unreliable storage, it is a main concern.
(Source: http://en.wikipedia.org/wiki/Cloud_computing)
INTRODUCTION TO CLOUD COMPUTING
Defining clouds:
Cloud computing is a paradigm that makes IT resources available
on demand over the Internet as subscription based service in a pay-
per-use model just like utility services such as electricity, gas,
telephone etc. Consumers need to pay only for the computing
services they use. In addition, consumers no longer need to invest
heavily in IT infrastructure or face difficulties in building and
maintaining complex IT resources. The IT resources mainly fall
into three categories:
(i) Application in the form of Software as a Service – SaaS
(ii) Environment in the form of Platform as a Service – PaaS
(iii) Infrastructure in the form of Infrastructure as a Service – IaaS.
INTRODUCTION TO CLOUD COMPUTING
Comparing Cloud Providers with Traditional IT
Service Providers:
Traditional IT Services:
(a) Provider operate the hardware, software, networks etc.
(b) Client pays for licensing fees for resources consumed.
(c) Provider operates the infrastructure in its own facilities.
(d) Client signs a long term contract/SLA with provider.
(Source: searchcloudcomputing.techtarget.com)
CLOUD PROVIDERS
#1: Amazon
For the second year in a row, the king of cloud is still Amazon Web Services. No other company has come close to the cloud-based innovation AWS provides.
Even Eli Lilly taking some of its business elsewhere ended up doing AWS a favor. Since that debacle over SLAs, Amazon has stepped up its support and now offers a premium
"white glove" service that routes your call to the nearest engineering specialist.
#2: Verizon/Terremark
Charging into the number two position on our list is Verizon. The telco giant had previously built its own cloud; high-quality stuff but with a commensurate price. The Four
Seasons of cloud, if you will: snooty service, small menu, long waits for a reservation and eye-watering bill. It was a test run, and apparently Verizon decided it needed some
expertise instead of re-inventing the wheel.
Verizon then bought Terremark, much as you or I would buy a coffee and a bagel. Not only is Terremark one of the premier Tier 1 hosters in the world, it's also a cloud supplier
to the coveted enterprise market, effectively moving Verizon into the top ranks.
Bigger than almost any competitor and with all the pipe in the world (literally), Verizon could be the King Kong of cloud. We’ll see if they can make it work or if Terremark
Cloud is doomed for post-acquisition mishandling.
#3: IBM
ew to the list is IBM with its Smart Business Test and Development Cloud. While Big Blue might be lugging a hundred years of IT baggage, it has finally launched
Infrastructure as a Service, although initially just for test and development purposes.
Despite its convoluted, muddled strategy in the cloud market, the Test and Dev service is winning enterprise business, which after all is IBM’s meat and potatoes. IBM
reportedly earned $30 million in cloud revenue last year; few others have the scale of the enterprise user base to ramp up that fast.
#4: Salesforce.com
Salesforce.com maintains a spot in the top five, thanks to its acquisition of Heroku. The company singlehandedly forced its way into the Platform as a Service market with this
buy, and it will give them legs to hit customers not interested in the patented "Salesforce.com maximum lock-in" feature offered on Force.com and their CRM platform.
The Software as a Service market is rapidly coalescing into a mature, well-defined space; props to Salesforce.com for grabbing on to something that keeps it relevant going
forward.
#5: CSC
Our second new entrant on the list is CSC. The IT integrator and service provider has cooked up an interesting private cloud service called BizCloud. The company will wheel
VCE -- the giant cloud-in-a-box system from VMware, Cisco and EMC -- into your IT shop. Ten weeks later, it will be integrated into all your messy, legacy IT systems,
turning on Infrastructure as a Service. CSC then manages your hardware; for extra capacity, you can hook into a public cloud service, also running on VCE.
CSC points to the trend of enterprises looking for practical ways to use (and get to) cloud computing; the company also performs massive-scale integrations with Google Apps
and other Software as a Service players. As a bridge to the cloud for many enterprises, CSC is on the front lines and on our top 10 list.
#6: Rackspace
CLOUD VENDORS
• 3Leaf Systems • 3PAR • 3Tera • 10Gen • Adaptivity
• Agathon Group • Akamai • Amazon EC2 • Apache Hadoop • Appirio
• Appistry • AppNexus • Apprenda • Appzero • Aptana
• Arjuna • Asankya • AT&T • CohesiveFT • Cordys
• Cumulux • Dataline • Dell • Desktoptwo • ElasticHosts
• Elastic Compute Cloud • Elastra • EMC • Engine Yard • ENKI
• Enomaly • Enomalism • Eucalyptus • eVapt • EyeOS
• FlexiScale • Intel • Interoute • iTricity • Joyent
• JumpBox • Juniper Networks • Kaavo • Kadient • Keynote Systems
• Layered Technologies • LinkedIn • LongJump • Meeza • Mezeo Software
• Microsoft • Morgan Stanley • MorphExchange • Netsuite • Rackspace
• Red Hat • Reservoir • Rhomobile • RightScale • Rollbase
• rPath • S3 • SalesForce.com • Savvis • ServePath/GoGrid
• SIMtone • Skytap • SLA@SOI • SmugMug • SOASTA
• StrikeIron IronCloud • Sun • Bluewolf • Boomi • Box-Net
• Booz Allen Hamilton • CA • Callidus Software • Cassatt • Cisco
• Citrix • Cloud9 Analytics • CloudBerry Lab • Cloudera • Cloudscale
• Cloudswitch • Cloudworks • Coghead • Force.com • Fortress ITX
• G.ho.st • GigaSpaces • GoGrid/ServPath • Google • gOS
• Grid Dynamics • Hadoop • Heroku • Hosting.com • HP
• Hyperic • IBM • iCloud • IMOD • newScale
• Ning • Nirvanix • Novell • OpenNebula • OpSource
• Oracle • OTOY • Parallels • ParaScale • Penguin Computing
• Platform Computing • Q-layer • Qrimp • Quantivo • Quickbase
• Terremark • The GridLayer • ThinkGrid • Unisys • Univa UD
• vCloud • Vertica • Virtual Workspaces • Vmware • WorkXpress
• Yahoo! • Zetta • Zimory • Zoho • Zuora
Source: http://cloudcomputing.sys-con.com/node/770174
INTRODUCTION TO CLOUD COMPUTING
CLOUD MODELS
According to article [1], cloud computing is divided among stack.
1. http://en.wikipedia.org/wiki/Cloud_computing, Feb-2010.
INTRODUCTION TO CLOUD COMPUTING
CLOUD MODELS
INTRODUCTION TO CLOUD COMPUTING
CLOUD MODELS
(…. As A Service) [Infrastructure, Software, Application, Platform,
Hardware, Social Networking, Desktop etc]
SaaS has its roots in an early kind of hosting operation carried out by
Application Service Providers (ASPs). The ASP business grew up soon
after the Internet began to mushroom, with some companies offering to
securely, privately host applications. Hosting of supply chain
applications and customer relationship management (CRM)
applications was particularly prominent, although some ASPs simply
specialized in running email. Prior to the advent of this type of service,
companies often spent huge amounts of money implementing and
customizing these applications to satisfy internal business
requirements. Many of these products weren’t only difficult to
implement but hard to learn and use. However, the most successful
vendors were those who recognized that an application delivered as a
service with a monthly fee based on the number of users had to be easy
to use and easy to stay with.
INTRODUCTION TO CLOUD COMPUTING
CLOUD MODELS: SAAS
Massively scaled Software as a Service
Environments such as Facebook, eBay, Skype, Google Apps, and
others are all designed for massive scaling. You may not think of many
of these Web sites as being software applications at all. Nevertheless,
all are used directly by businesses, for business purposes. For example,
some companies use the social networking site Facebook as a free
intranet for its employees. Online auctioneer eBay is the basis of more
than 500,000 small businesses, Skype (free online calls and video) is
used by small businesses the world over, and Google Apps (messaging
and collaboration tools) has over a million different businesses
enrolled.
INTRODUCTION TO CLOUD COMPUTING
CLOUD MODELS: SAAS
Massively scaled Software as a Service
Economies of scale
The companies that provide massively scaled Software as a Service
achieve dramatic economies of scale — cost efficiencies gained from
reducing per-unit costs when more of the same item is produced or
more of the same workloads are processed.
It’s worth listing all the reasons why:
✓ The standardized workloads can be executed on a highly integrated, massively replicable infrastructure
stack. They don’t have to support a wide array of workloads and a heterogeneous stack of hardware,
middleware, OS, and so on.
✓ The computer hardware and network is highly streamlined and can be bought in bulk and configured to
allow expansion. Often these companies require that hardware be engineered for their unique scaling
requirements.
✓ All software can be stripped down so that only what is necessary is loaded.
✓ The service/software itself is written from scratch in a cloud-optimized way, tailored for efficiency at an
instruction level.
✓ The provider may not offer or guarantee a specific service level.
✓ There is no need for virtualization technology to build virtual machines. The software can be engineered to
the bare metal.
✓ The profile of the workload is measurable and predictable simply by numbers of users.
INTRODUCTION TO CLOUD COMPUTING
Getting Inside the Cloud
• Technical Interface
INTRODUCTION TO CLOUD COMPUTING
• Cloud Resources
NATURE OF CLOUD
• Traditional data center
• Cost of cloud data center
• Scaling computer systems
• Economics
• Cloud Work Load
• Managing data on cloud
• Public, Private and Hybrid Cloud
NATURE OF CLOUD
• Traditional data center
NATURE OF CLOUD
• Cost of cloud data center
NATURE OF CLOUD
• Scaling computer systems
NATURE OF CLOUD
• Economics
NATURE OF CLOUD
• Cloud Work Load
NATURE OF CLOUD
• Managing data on cloud
NATURE OF CLOUD
• Public, Private and Hybrid Cloud
NATURE OF CLOUD
• Public, Private and Hybrid Cloud
Cloud Computing Deployment Models
CLOUD ELEMENTS
• Infrastructure as a Service
• Platform as a Service
• Software as a Service
CLOUD ELEMENTS
• Infrastructure as a Service
CLOUD ELEMENTS
• Platform as a Service
CLOUD ELEMENTS
• Software as a Service
Systems
Characteristics
Cluster Grid Cloud
Commodity computers and high-end servers `and network
Population Commodity computers High-end computers (servers, clusters)
attached storage
Size/scalability 100s 1000s 100s to 1000s
One of the standard OSs (Linux,
Node Operating System (OS) Any standard OS (dominated by Unix) A hypervisor (VM) on which multiple OSs run
Windows)
Ownership Single Multiple Single
Interconnection Dedicated, high-end with low
Mostly Internet with high latency and low bandwidth Dedicated, high-end with low latency and high bandwidth
Network speed latency and high bandwidth
Traditional login/password-based. Public/private key pair based authentication and Each user/application is provided with a virtual machine.
Security/privacy Medium level of privacy depends on mapping a user to an account. Limited support for High security/privacy is guaranteed. Support for setting
user privileges. privacy. per-file access control list (ACL).
Failure management Limited (often failed Limited (often failed tasks/applications are Strong support for failover and content replication. VMs
(Self-healing) tasks/applications are restarted). restarted). can be easily migrated from one node to other.
Multi-clustering within an Limited adoption, but being explored through High potential, third party solution providers can loosely
Internetworking
Organization research efforts such as Gridbus InterGrid tie together services of different Clouds
Science, business, enterprise Collaborative scientific and high throughput Dynamically provisioned legacy and web applications,
Application drivers
computing, data centers computing applications Content delivery