0% found this document useful (0 votes)
258 views184 pages

Cloud Computing

Uploaded by

Mayuri Falke
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
258 views184 pages

Cloud Computing

Uploaded by

Mayuri Falke
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 184

SUBJECT CODE : 310254(C)

As per Revised Syllabus of


SAVITRIBAI PHULE PUNE UNIVERSITY
Choice Based Credit System (CBCS)
TE. (Computer) Semester - VI Elective - I

CLouD COMPUTING

Iresh A. Dhotre
M.E. (Information Technology)
Ex-Faculty, Sinhgad College of Engineering,
Pune.

= "TECHNICAL
I~
N=
~~ P
~~ |
SINCE 1993 An Up-Thrust for Knowledge

U]
CLOUD COMPUTING

Subject Code : 310254(C)

TE. (Computer Engineering) Semester - VI Elective - lI

© Copyright with Author


All publishing rights (printed and ebook version) reserved with Technical Publications. No part of this book
should be reproduced in any form, Electronic, Mechanical, Photocopy or any information storage and
retrieval system without prior permission in writing, from Technical Publications, Pune.

Published by :
="TECHNICAL Amit Residency, Office No.1, 412, Shaniwar Peth,
S PUBLICATIONS Pune - 411030, M.S. INDIA, Ph.: +91-020-24495496/97
sce 255 An Up-Thrust for Knowledge Email : [email protected] Website : www.technicalpublications.org

Printer :
Yogiraj Printers & Binders
Sr.No. 10/1A,
Ghule Industrial Estate, Nanded Village Road,
Tal. - Haveli, Dist. - Pune - 411041.

ISBN 978-93-91567-40-8

939115674 SPPU 19

9789391567408 [1] [{/]


PREFACE
The importance of Cloud Computing is well known in various engineering fields.
Overwhelming response to my books on various subjects inspired me to write this book. The
book is structured to cover the key aspects of the subject Cloud Computing.

The book uses plain, lucid language to explain fundamentals of this subject. The book
provides logical method of explaining various complicated concepts and stepwise methods
to explain the important topics. Each chapter is well supported with necessary illustrations,
practical examples and solved problems. All the chapters in the book are arranged in a
proper sequence that permits each topic to build upon earlier studies. All care has been
taken to make students comfortable in understanding the basic concepts of the subject.

Representative questions have been added at the end of each section to help the
students in picking important points from that section.

The book not only covers the entire scope of the subject but explains the philosophy of
the subject. This makes the understanding of this subject more clear and makes it more
interesting. The book will be very useful not only to the students but also to the subject
teachers. The students have to omit nothing and possibly have to cover nothing more.

[ wish to express my profound thanks to all those who helped in making this book a
reality. Much needed moral support and encouragement is provided on numerous
occasions by my whole family. | wish to thank the Publisher and the entire team of
Technical Publications who have taken immense pain to get this book in time with quality
printing.

Any suggestion for the improvement of the book will be acknowledged and well
appreciated.

Author
D, A, Dhotre

Dedicated to God.

(iid)
SYLLABUS
Cloud Computing - 310254(C)
Credit : Examination Scheme :

Mid-Semester (TH) : 30 Marks


03 End-Semester (TH) : 70 Marks

Unit] Introduction to Cloud Computing


Importance of Cloud Computing, Characteristics, Pros and Cons of Cloud Computing, Migrating
into the Cloud, Seven-step model of migration into a Cloud, Trends in Computing. Cloud Service
Models : SaaS, PaaS, laaS, Storage. Cloud Architecture : Cloud Computing Logical Architecture,
Developing Holistic Cloud Computing Reference Model, Cloud System Architecture, Cloud
Deployment Models. (Chapter - 1)

UnitIl Data Storage and Cloud Computing


Data Storage : Introduction to Enterprise Data Storage, Direct Attached Storage, Storage Area
Network, Network Attached Storage, Data Storage Management, File System, Cloud Data Stores,
Using Grids for Data Storage. Cloud Storage : Data Management, Provisioning Cloud storage, Data
Intensive Technologies for Cloud Computing. Cloud Storage from LANs to WANs : Cloud
Characteristics, Distributed Data Storage. (Chapter - 2)

UnitIll Virtualization in Cloud Computing


Introduction : Definition of Virtualization, Adopting Virtualization, Types of Virtualization,
Virtualization Architecture and Software, Virtual Clustering, Virtualization Application, Pitfalls of
Virtualization. Grid, Cloud and Virtualization : Virtualization in Grid, Virtualization in Cloud,
Virtualization and Cloud Security. Virtualization and Cloud Computing : Anatomy of Cloud
Infrastructure, Virtual infrastructures, CPU Virtualization, Network and Storage Virtualization.
(Chapter
- 3)

Unit IV Cloud Platforms and Cloud Applications


Amazon Web Services (AWS) : Amazon Web Services and Components, Amazon Simple DB, Elastic
Cloud Computing (EC2), Amazon Storage System, Amazon Database services (Dynamo DB).
Microsoft Cloud Services : Azure core concepts, SQL Azure, Windows Azure Platform Appliance.
Cloud Computing Applications : Healthcare : ECG Analysis in the Cloud, Biology : Protein
Structure Prediction, Geosciences : Satellite Image Processing, Business and Consumer Applications
: CRM and ERP, Social Networking, Google Cloud Application : Google App Engine. Overview of
OpensStack architecture. (Chapter - 4)

(iv)
UnitV Security in Cloud Computing
Risks in Cloud Computing : Risk Management, Enterprise-Wide Risk Management, Types of Risks
in Cloud Computing. Data Security in Cloud : Security Issues, Challenges, advantages,
Disadvantages, Cloud Digital persona and Data security, Content Level Security. Cloud Security
Services : Confidentiality, Integrity and Availability, Security Authorization Challenges in the Cloud,
Secure Cloud Software Requirements, Secure Cloud Software Testing. (Chapter - 5)

Unit VI Advanced Techniques in Cloud Computing


Future Trends in cloud Computing, Mobile Cloud, Automatic Cloud Computing : Comet Cloud.
Multimedia Cloud : IPTV, Energy Aware Cloud Computing, Jungle Computing, Distributed Cloud
Computing Vs Edge Computing, Containers, Docker, and Kubernetes, Introduction to DevOps. 10T
and Cloud Convergence : The Cloud and loT in your Home, The IOT and cloud in your
Automobile, PERSONAL : [oT in Healthcare. (Chapter - 6)


TABLE OF CONTENTS

Chapter - 1 Introduction to Cloud Computing (1-1)to (1-34)

1.1 Overview of Cloud COMPUEING....ccceiiiiiei cieseee reese 1-2


1.1.1 Cloud COMPONENLES c..eeevieeiieies steers seas bese sae ses sees 1-3

1.2 Importance of Cloud COMPULING.....cccceiiiiie ieee eeee ns 1-4


1.2.1 Characteristics of Cloud COmMPULING .....eeveeiieree ieee 1-5

1.2.2 Pros and Cons of Cloud COMPULING ......eevieiiirer


eee eee eee 1-5

1.2.3 Cloud APPlICAtiONS ....ooveeeeeeiee cere eee eee ee 1-6

1.3 Migrating into the Cloud...


eee eens 1-7
1.4 Seven-step Model of Migration into a Cloud.......cccccceeeciiei ices ee 1-9

1.5 Trends iN COMPUEINEG....coiiiieiicieeeciee


creer essere esters ese sere e ee snre ee enee 1-10
1.6 Cloud Service MOIS. .... coor
eee eee eee eas 1-11
1.6.1 Software as a Service (SaaS) ...coiivieueeeee
cece eee eee eee 1-12

1.6.2 Platform as a Service (PaaS)........cooeueevieeiiereeieeeeeeee


eee eevee eres 1-13

1.6.3 Infrastructure as a Service (18aS) .....cocevveeieeeeiee cece 1-15

1.6.4 Difference between 1aas, PaaS and SaaS........coccveeevvveeevieeee cece 1-16

1.7 Cloud Computing Logical Architecture .......cccce eevee ccee ccc cceeeceeeeceee 1-17
1.7.1 Cloud Computing Service Architecture.......ccooveeiieeee eine 1-19

1.8 Developing Holistic Cloud Computing Reference Model........ccccceecveernnnennn. 1-20


1.8.1 Complete View of Cloud Management ........coceevevierieeneneeeeeenee
eee 1-21

1.8.2 Cloud Computing Reference Model ..........ooceeieieireeiieneeeeee


ee, 1-22

1.9 Cloud System Archit@CtUIE .....cooccieei ieee cee ec ee es rree ee eeesree ee sreeeeaee 1-24
1.9.1 Single Cloud Site ArchiteCtUres......coceevieeiiireeieee eee 1-25

1.9.2 Redundant 3-Tier Architecture .......oeeev reece 1-25

1.9.3 Multi-Data Centre ArchiteCture......ooeeeeeeeeeireee eee 1-26

1.10 Cloud Deployment MOAEIS .....oei ieee ccc 1-27


1.10.1 Difference between Public and Private Cloud........cccoeeeeerveeneieeceeecneeens 1-30

1.11 Multiple Choice Questions With ANSWErS .......ccccevcieer icine eves 1-30

Chapter - 2 Data Storage and Cloud Computing (2-1) to (2 - 28)

2.1 Introduction to Enterprise Data STOrage .....cccvvveeevecieeeicieeeeciee


secre ese enes 2-2
2.1.1 Direct Attached StOFrage ......ooveveeeeiiieeee eee ees 2-2

2.1.2 Storage Area NEEWOIK ....cooceeeeiieeeeee cee es 2-3

2.1.3 Network - attached Storage ......cceeeeeeree eee 2-5

2.1.4 Comparison between NAS and SAN .......ccorr ieee eee 2-9

2.1.5 Comparison between DAS, NAS and SAN .......coccoviirinienes sees 2-10

2.2 Data Storage Management. ...c.eei ieee ccciieeie cece e essere reese esr e eee 2-10
2.2.1 Data Storage Management TOOIS .......c.coeveireeiieneneee
eee eee 2-10

2.2.2 Storage Management ProCESS......oocevveeeereereese


ieee sees sees es 2-11

2.2.3 Cloud Provisioning .....ccccveerereieiesiieste


essere esses see seas ees 2-11

0 J TLR V3 Yo SR 2-12
2.3.1 FAT FIle SYSEEM coon eae sae 2-13

2.3.2 NTFS cee


see sees ees sees see eee s 2-13
2.3.3 Cloud File SYStEM.. ecoeeeeeee ses 2-13

2.4 CloUd Data StOMES.... ii ieieiueieieeeiee site ree erica ete setae see e sues ete eseeesaee sabes see e anne enns 2-15
2.4.1 Distributed Data Store ......cceveee eerie 2-16

2.4.2 Data Store TYPES .coeeiee eee eerste


eres sess ee sree sean ssn sees 2-16

2.4.2.1 BIgTABI cc. iiiete ste staat ste eae anata ata een 2-16

2.4.2.2 DYNAIMO cities


ete eet atte testes atte eae t eet beta nae tae eate tenant bee ataan 2-17

2.5 Using Grids for Data StOrage....c..cccevriiiiieicieeeecieesesiee


esse es esaeessnre es snaree sens 2-18

BN I © To TU To JR) Lo] 1 = JSR 2-19


2.6.1 Data Management for Cloud Storage........ccoevevieerineeeries
scree eeee eee 2-20

2.6.2 Cloud Data Management INtErface ........ccceveeeverrenee sees 2-21

2.6.3 Cloud Storage REQUIrEMENL .......cevereiiriiiriieeee eee ses 2-21

2.6.4 Provisioning Cloud StOrage........cceverreeeiie


eset eres 2-22
2.6.5 Data Intensive Technologies for Cloud Computing .........cceeeeeeereereneeene 2-23

2.7 Cloud Storage from LANS t0 WANS .......ccciiiriiieeccirsieeeecee


serene esses seeneee ene 2-24
2.7.1 Cloud Characteristics .......eevreererersereeesese essere eens 2-24

2.7.2 Distributed Data Storage ....ccoceeeveereereeeiesee


eee eee 2-24

2.8 Multiple Choice Questions With ANSWENS .......ccccieeiiceeecciie


eerie ees 2-26

Chapter - 3 Virtualization in Cloud Computing (3-1) to (3-30)

3.1 Definition of Virtualization ..........cccoeiriiiii eee 3-2


3.2 Adopting Virtualization .....coccceee ieee cers sere eee esr 3-3

3.3 Types of VIrtualizatioNns.....cccccceeiiciee sciences seers eserves eens ene 3-4


3.3.1 EXAMPLE t WINE ceeeee es 3-5

3.3.2 Server Virtualization .....eeeeeee eee es 3-6

3.3.3 Operating Level Virtualization ...........ooeeeieiin iene, 3-9

3.3.4 Para-Virtualization ......ceeeeeeereeeeeeeeee eee 3-11

3.4 Full VIrtUAlIZatioNn ...oooeeeoce ccc eee eee ees 3-12


3.4.1 Memory Virtualization .......oceeevee ieee 3-13

3.4.2 I/O VIrtUlIZALION. ...oeee cece


sees seas 3-14

3.4.3 Difference between Full and Para Virtualization..........cceveeeveereeerennenens 3-15

3.5 Storage Virtualization ......cee cece ccc ceeserre eer ee ee ee ee ee ee eee 3-15
3.5.1 Network Virtualization .........ceeeeeeeeieeeeeireeeee cee 3-18

3.6 Virtual ClUSLEIING eee iceeeciee ects ects estes stresses estes ea reese re ese rare essere ee enes 3-19
3.7 Virtualization Application......cceeeiciir ices ccee ec eeeeseeeeeeeeeee 3-20

3.8 Pitfalls of Virtualization ........coocee iirc 3-21


3.9 Grid, Cloud and Virtualization .........coeccueeeeee
ieee eee eases 3-21
3.9.1 Virtualization in Grid.....o.eeeeeereeeeee eee se 3-21

3.9.2 Virtualization in Cloud .....c.eoeereeeieeeeeeeee eee se 3-22

3.9.3 Difference between Cloud and Virtualization .........ccoceeeveeeeeenseee


serene 3-23

3.9.4 Virtualization and Cloud SECUFItY .....ceeevieriireeiieeeeee eee, 3-23

(vii)
3.10 Virtualization and Cloud COMPUTING ......coecciereiiier reciever sees eens 3-25
3.10.1 Anatomy of Cloud INfrastructure........coceeveeeeiie ieee 3-25

3.10.2 Virtual INfrastrUuCtures ...oeevveeeeeieeee sees see ees 3-26

3.10.3 CPU ViIrtualization......cccecerveee seers seeeeeeeeeeee


ese s 3-26
3.10.4 Network and Storage Virtualization ...........ccceeeeeeeniineeee eee 3-27

3.11 Multiple Choice Questions With ANSWEIS ........cccceeiivieeiciee


ec ciee eccrine eens 3-28

Unit - IV

Chapter - 4 Cloud Platforms and Cloud Applications (4 -1) to (4 - 32)

4.1 AMAzon WED SEIVICES ...co ities


ese eae ete ene eens 4-2
4.1.1 COMPONENES eee reeset seers ses eres see sass see sess snes see sne se saes 4-3

4.1.2 Advantages and Disadvantages of AWS ......cccooirinn eee, 4-4

4.2 Elastic Cloud COMPUEING c.cuveveiiieie ices csi crie eee e esses erase esas senna e anes 4-4
4.2.1 Configuring Amazon EC2 LINUX INStANCES ....coverierierer cies eee 4-6

4.2.2 AMAZON SI oii es 4-10

4.23 Amazon CloudWatch o.oo 4-12

4.3 Amazon STOrage SYSTEM uve essere essere reese errr essere eee eens 4-13
4.4 Amazon Database SErvICeS ......ccuiii iii ieee erie eee ese eee eee eee 4-14
4.41 DYNAMO DB eee eee 4-15

4.4.2 Difference between DynamoDB and Amazon S3 ........coccevveveienen sence. 4-16

4.5 Microsoft Cloud Services : AZUIe ......occeeieieieiie


creases eee ees 4-17

4.6 Cloud Computing Applications. ....ccccceiiciee


cere cries reese 4-19
4.6.1 Healthcare : ECG Analysis in the Cloud........cocooveriinee ieee 4-20

4.6.2 Biology : Protein Structure Prediction .........cceeeeveeerer ieee 4-21

4.6.3 Geosciences : Satellite Image Processing ........cooeevveeveeeieneneese sree 4-22

4.6.4 Business and Consumer Applications : CRM and ERP ........ccocceeievrenennnne 4-23

4.6.5 SoCial NEtWOIKING ...eeveeeiie cereeee 4-24

4.7 Google Cloud Application : Google App ENgine.....cccceeeeeevcieeecciee cece, 4-25


4.8 Overview of OpenStack Architecture ......cccocceeeccee ccc e, 4-28

4.9 Multiple Choice Questions With ANSWEIS.......ceeevceeeiiciee


erie eerie sree 4-29

(ix)
Chapter - 5 Security in Cloud Computing (5-1) to (5-14)

5.1 Risks in Cloud COMPULING....cieiiiiireiiier eters scieressee essere essere seers se ree essen snes 5-2
5.1.1 Risk Management ........ooieeieiin ieee cee eee ee eee 5-2

5.2 Enterprise-wide Risk Management.......ccccevicieeeniee


esis cries cries erie serene 5-3

5.3 Types of Risks in Cloud COMPULING .....ceeviiieiiiiie


cries csr eres 5-4
5.4 Data Security in Cloud : Security Issues and Challenges......cccccocevvcieeeiieenn. 5-6
5.4.1 AdVANTAEES .ooveeieeeeeeee eee eee eee eee sheen 5-7

5.4.2 DiSAVANTAZES....ceervieeiiriiriieeie esters ee eee 5-7

5.5 Content LEVEl SECUILY .ueviiuieeeicieeecteeeceeeerteeeseeeeseeeeseeseeeesnr


eee senna eens 5-8
5.6 Cloud SECUITY SEIVICES .uviiiiiieiiiieeeciee
ects esters esters esters esters estes seen e sense ee ennes 5-8

5.7 Security Authorization Challenges in the Cloud.........ccccvrviiiivce nce, 5-10

5.8 Secure Cloud Software ReqUIremMeENtS .....c.eevceeeeiiceeeeccieeeeciee


sees sees 5-11
5.9 Secure Cloud Software TESTING ...c.cccveericiiie
ieee eer eereeeeeeeesee ea 5-11
5.9.1 Type of Testing in Cloud .........coeiieiiriiieie ees 5-13

5.9.2 Benefit of Cloud-based TESHING ......cocveveeierer eee 5-13

5.10 Multiple Choice Questions With ANSWEIS ........cccceeiiceee


cries ccee ees enes 5-14

Unit - VI

Chapter - 6 Advanced Techniques in Cloud Computing


(6-1) to (6-30)

6.1 Future Trends in Cloud COMPUTING c.uevrieericiie eccrine eens 6-2


6.1.1 Just Enough Operating System ........oocvuerieiiiies eee 6-2

(SA Vo]
o 11 T= [TH oR 6-3
6.2.1 Advantages and Disadvantages .........cceeeevvierieireniiesee eee 6-5

6.2.2 Mobile Cloud Application ......cocee viiiee 6-5

6.3 Automatic Cloud COMPUEING......ccv ieee icici eres ees 6-6


6.3.1 COMELCIOU covets
see eee eee sbeebs 6-8

6.4 Multimedia CloUd........c.oo ieee eee eee eas 6-9


6.4.1 IPTV cee a es 6-11

6.5 Energy Aware Cloud COMPUEING ...cooiiiirriiiiie


cece eee 6-12
6.5.1 Gren ClOU ..ooeeeeeeeeeie eee
sees sree sees 6-12

6.5.2 Green COMPULING cooeeeeeeeeeee eee eee sees ssn 6-13

6.5.3 Energy-Saving Software Techniques.........ccoeeeiiinieeee ees 6-14

6.5.4 Key Issue Related to Cloud Computing Energy Efficiency ........ccooveveenenne 6-15

6.6 JUNEIE COMPULING..c.eeiiieeeecieeeccieeeccree


ec eee ese essere esters eee esters sere eee enree eens 6-16

A 0 To Tol C=] URS 6-17


6.7.1 Process SImplification .......ccoeceeveiiieiieceeee ees 6-17

6.7.2 Broad Support and Adoption .........cecee eerie 6-18

6.7.3 ArChitECtUNE eee se ns 6-19

6.7.4 Container and KUDEIMELES ........ocevvvereeeeireee cece 6-20

6.8 INTroduction t0 DEVOPS ...ueeieicieee cesses crtre ee ertte essere erates ses eennre es senre ee enes 6-23

6.9 10T and Cloud CONVEIENCE. ......eii ieee ecieeeeciee essere essere eeenre ese snses assess snnreeeenes 6-23
6.9.1 The Cloud and 10T in yoUr HOME ....cc.coiiiiiiirieiieee sees 6-23

6.9.1.1 SMArt LIGRTING «o.oo


cies ste cesta ste steer at aba sae senasensaens 6-24

6.9.2 The IOT and Cloud in your Automobile .........cocovvvirieiini eee 6-25

6.9.2.1 Remote Vehicle Diagnostics... couture es cerieiieie cs aiestieie es aeestieesssassensaens 6-26

6.9.3 PERSONAL: [OT in HEAIthCare. ....coeveeee eerie 6-27

6.10 Multiple Choice Questions With ANSWENS ........ccccveiicieeicieeecciee


eccrine eens 6-28

Solved Model Question Papers M-1)to (M-4)


Notes

(xii)
UNIT - 1

Introduction to Cloud Computing

Syllabus
Importance of Cloud Computing, Characteristics, Pros and Cons of Cloud Computing, Migrating
into the Cloud, Seven-step model of migration into a Cloud, Trends in Computing. Cloud Service
Models : SaaS, PaaS, laa, Storage. Cloud Architecture : Cloud Computing Logical Architecture,
Developing Holistic Cloud Computing Reference Model, Cloud System Architecture, Cloud
Deployment Models.

Contents
1.1 Overview of Cloud Computing ............. March-19, ------------- Marks 5
1.2 Importance of Cloud Computing. . .......... March-19, ------------- Marks 5
1.3 Migrating into the Cloud
14 Seven-step Model of Migration into a Cloud
1.5 Trends in Computing
1.6 Cloud Service Models .................. March-19, ------------- Marks 4
1.7 Cloud Computing Logical Architecture
1.8 Developing Holistic Cloud Computing Reference Model
1.9 Cloud System Architecture
1.10 Cloud Deployment Models . ............... March-19, ------------- Marks 6
1.11 Multiple Choice Questions

(1-1)
Cloud Computing 1-2 Introduction to Cloud Computing

EEN overview of Cloud Computing


Idea of cloud computing was introduced by computer scientist John McCarthy
publicly in 1961.
Then in 1969, Leonard Kleinrock, a chief scientist of the ARPANET project
comments about Internet.
The general public has been leveraging forms of Internet-based computer utilities
since the mid-1990s through various incarnations of search engines, e-mail services,
open publishing platforms and other types of social media.
Though consumer-centric, these services popularized and validated core concepts
that form the basis of modern-day cloud computing.
The Salesforce.com provides remote service from 1990 to organization. Amazon
launched its web services in 2002 and it provides services to organization for
storage and remote computing.
Cloud computing definition as per Gartner "a style of computing in which scalable
and elastic IT-enabled capabilities are delivered as a service to external customers
using Internet technologies".
In 2008, Gartner's original definition of cloud was changed. In the definition,
"massively scalable” was used instead of "scalable and elastic.”
NIST definition of cloud : Cloud computing is a pay-per-use model for enabling
available, convenient, on-demand network access to a shared pool of configurable
computing resources (e.g., networks, servers, storage, applications, services) that
can be rapidly provisioned and released with minimal management effort or
service-provider interaction.
The above cloud definition was published by NIST in 2009, followed by a revised
version after further review and industry input that was published in September
of 2011.
Cloud computing refer to a variety of services available over the Internet that
deliver compute functionality on the service provider's infrastructure.
= Its environment (infrastructure) may actually be hosted on either a grid or
utility computing environment, but that doesn't matter to a service user.
= Cloud computing refer to a variety of services available over the Internet that
deliver compute functionality on the service provider's infrastructure. Its
environment (infrastructure) may actually be hosted on either a grid or utility
computing environment, but that doesn't matter to a service user.
Cloud Computing = Software as a service + Platform as a service
+ Infrastructure as a service + Data as a service

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 1-3 Introduction to Cloud Computing

¢ Cloud computing is a general term used to describe a new class of network based
computing that takes place over the Internet, basically a step on from utility
computing.
e In other words, this is a collection/group of integrated and networked hardware,
software and Internet infrastructure (called a platform).
e Fig. 1.1.1 shows cloud symbol. It
denotes cloud boundary.
e Using the Internet for
communication and transport
provides hardware, software and
networking services to clients.
e These platforms hide the complexity
and details of the underlying
infrastructure from users and
applications by providing very
simple graphical interface or APL Fig. 1.1.1 Cloud symbol
e In addition, the platform provides on demand services that are always on
anywhere, anytime and anyplace. Pay for use and as needed, elastic.
e The hardware and software services are available to the general public, enterprises,
corporations and business markets.

EEEN Cloud Components


¢ Cloud computing solutions are made up of several elements. Fig. 1.1.2 shows
cloud components.
1. Clients : Mobile, terminals or regular computers.
2. Benefits : Lower hardware costs, lower IT costs, security, data security, less
power consumption, ease of repair or replacement, less noise.
3. Data centers : Collection of servers where the application to subscribe is
housed. It could be a large room in the basement of your building or a room
full of servers on the other side of the world
4. Virtualizing servers : Software can be installed allowing multiple instances of
virtual servers to be used and a dozen virtual servers can run on one physical
server.
5. Distributed servers : Servers don't all have to be housed in the same location.
It can be in geographically disparate locations. If something were to happen at
one site, causing a failure, the service would still be accessed through another
site. If the cloud needs more hardware, they can add them at another site.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 1-4 Introduction to Cloud Computing

\\
Wy
My
\\
Data centers

Client computers
\\

\\
Distributed servers

Fig. 1.1.2 Cloud components

1. What is cloud computing ? What are the components of cloud computing ? Explain with neat
diagram. SPPU : March-19, In Sem, Marks 5

ER importance of Cloud Computing


It is clear that utilizing the cloud is a trend that continues to grow. It provides
better data storage, data security, flexibility, increased collaboration between
employees, and changes the workflow of small businesses and large enterprises to
help them make better decisions while decreasing costs.
The significance of the cloud is increasing exponentially. Gartner forecasts that the
cloud services market will grow 17.3 % in 2019 ($206.2 billion) and by 2022, 90 %
of organizations will be using cloud services.
Considering all the potential and development cloud computing has undergone in
recent years, there are also many challenges businesses are facing.
Cloud storage works by storing data on remote servers, where it can be
maintained, managed, backed up and accessed remotely. Data stored in the cloud
is accessible by any device at any time, as long as permissions are in place.
Despite its accessibility, data stored via the cloud is extremely safe and secure.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 1-5 Introduction to Cloud Computing

Cloud computing offers the media streaming companies a cost effective, flexible,
secure and scalable solution for setting up their work-flow. This serves to meet the
media streaming companies requirements for both video on demand and live
streaming services.
During a live broadcast of India v/s Pakistan match, the traffic is easily 10-20
times that on normal days. Provisioning for servers and other infrastructure would
end up creating a large overlay of upfront cost for infrastructure which otherwise
would not be utilised in normal days. The cloud with its pay per use strategy
offers economies of scale with effective costing strategies at a very high rate of
reliability and security.

1.2.1 | Characteristics of Cloud Computing

1. On-demand self-service : A consumer can unilaterally provision computing


capabilities, such as server time and network storage, as needed without requiring
human interaction with each service's provider.
. Ubiquitous network access : Capabilities are available over the network and
accessed through standard mechanisms that promote use by heterogeneous thin or
thick client platforms.
. Location-independent resource pooling : The provider's computing resources are
pooled to serve all consumers using a multi-tenant model, with different physical
and virtual resources dynamically assigned and reassigned according to consumer
demand.
. Rapid elasticity : Capabilities can be rapidly and elastically provisioned to quickly
scale up, and rapidly released to quickly scale down.
. Pay per use : Capabilities are charged using a metered, fee-for-service, or
advertising-based billing model to promote optimization of resource use.

EEX] Pros and Cons of Cloud Computing


Pros of cloud computing :
1. Lower computer costs : Since applications run in the cloud, not on the desktop
PC, your desktop PC does not need the processing power or hard disk space
demanded by traditional desktop software.
2, Improved performance : Computers in a cloud computing system boot and run
faster because they have fewer programs and processes loaded into memory.
3. Reduced software costs : Instead of purchasing expensive software applications,
you can get most of what you need for free.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 1-6 Introduction to Cloud Computing

4. Instant software updates : When you access a web-based application, you get the
latest version - without needing to pay for or download an upgrade.
5. Improved document format compatibility : You do not have to worry about the
documents you create on your machine being compatible with other user's
applications or operating systems.
6. Unlimited storage capacity : Cloud computing offers virtually limitless storage.
7. Increased data reliability : Unlike desktop computing, in which if a hard disk
crashes and destroy all your valuable data, a computer crashing in the cloud
should not affect the storage of your data.
8. Universal document access : All your documents are instantly available from
wherever you are.
9. Latest version availability : The cloud always hosts the latest version of your
documents; as long as you are connected, you are not in danger of having an
outdated version.
10. Easier group collaboration : Sharing documents leads directly to better
collaboration.
11. Device independence : Move to a portable device and your applications and
documents are still available.

Cons of cloud computing :


1. It requires a constant Internet connection : Cloud computing is impossible if you
cannot connect to the Internet.
2. Features might be limited.
3. Stored data might not be secure : With cloud computing, all your data is stored
on the cloud.
4. Does not work well with low-speed connections.

EEX] Cloud Applications


1. Through cloud cost flexibility, online marketplace gains access to more powerful
analytics online. Cloud takes away the need to fund the building of hardware,
installing software or paying dedicated software license fees.
2. Greater business scalability enables online video retailer to meet spikes in
demand : Cloud enables businesses not just IT operations to add or provision
computing resources just at the time they're needed.
3. Greater market adaptability provides online entertainment platform the ability to
reach any type of customer device. A third of the executives we surveyed believe

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 1-7 Introduction to Cloud Computing

cloud can help them adapt to diverse user groups with a diverse assortment of
devices.
4. Masked complexity enables access to services, no matter how intricate the
technology they're built on.
5. With context-driven variability, "intelligent assistants” are possible. "Because of its
expanded computing power and capacity, cloud can store information about user
preferences, which can enable product or service customization,” the report states.
6. Ecosystem connectivity enables information exchange across business partners.

1. What are the advantages and disadvantages of cloud computing ?


SPPU : March-19, In Sem, Marks 5

EEX Migrating into the Cloud


Cloud migration is the process of transferring data, application code and other
technology-related business processes from an on-premise or legacy infrastructure
to the cloud environment.
Cloud migration is a phenomenal transformation in the business information
system domain as it provides adequate services for the growing needs of
businesses. However, moving data to the cloud requires preparation and planning
in deciding on an approach.
Technology companies providing cloud and managed services should work closely
with businesses and support them on three main aspects as part of their cloud
migration journey :
a) Developing the right approach by ensuring clarity on the desired business
outcome or end result.
b) Taking advantage of relevant talent and expertise to address companie’s core
business issues.
c) Using appropriate toolsets that are user-friendly, secure and viable in the long
term.
Your cloud computing migration strategy must deliver elasticity, agility, and
scalability to help your organization take advantage of emerging opportunities and
pivot to address business and industry changes.
Achieving these goals will broadly employ a mix of Software as a Service (SaaS),
Infrastructure as a Service (IaaS) and Platform as a Service (PaaS) cloud

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 1-8 Introduction to Cloud Computing

architecture layers. You may employ these across hybrid and multi-cloud
environments based on business needs and best practices.
¢ Cloud migration strategy should focus on four broad areas :
a) Security that determines application and data access controls, data security at
rest and in transit and backup and data recovery.
b) Governance and compliance that makes sure you meet regulatory compliance
and all aspects of data security to go beyond the limits of shared security
models of cloud services providers.
¢) Cost management that looks at immediate and long-term cost savings that
guide where and how you migrate, manage, and monitor applications and
workloads.
d) Accessibility, resilience, and scalability that are driven by customer and
workforce UX needs, service enhancements and innovation.
e Benefits of migrating to the cloud include :
a) Increased agility and flexibility
b) Ability to innovate faster
c) Easing of increasing resource demands
d) Better managing of increased customer expectations
e) Reduction in costs
f) Deliver immediate business results
g) Simplify IT
h) Shift to everything as-a-service
i) Better consumption management
j) Cloud scalability
k) Improved performance
Cloud migration process :
e Cloud migration process is divided into three phases : plan, execute (run) and
monitor
1. Plan : Cloud migration requires solid planning to be successful. It define following
parameters :
a. Identify business objective
b. Identify key business drivers
c. Get executive sponsorship
d. Providing full visibility into your on-premise environment, including all system
dependencies.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 1-9 Introduction to Cloud Computing

f. Evaluating performance, server, and security requirements.


2. Execute or run : Once your environment has been assessed and a plan has been
mapped out, it's necessary to execute your migration. The main challenge here is
carrying out your migration with minimal disruption to normal operation, at the
lowest cost, and over the shortest period of time.
3. Monitor : After a successful migration, tools and processes should be implemented
to monitor the new cloud environment.
eo A formal issue tracking process should be created to ensure that everyone
impacted by the migration has a way to report problems, and IT has a simple way
to manage all requests.
Challenges of cloud migration :
1. Interoperability : It is no easy feat to get existing applications to communicate
with newer cloud environments.
2. Resource availability : The migration process might require taking in-house
servers temporarily offline. But downtime could be disastrous to application
performance, thus customer loyalty, if not supported by a proper plan for disaster
recovery.
3. Data integrity : How will user keep data secure while moving it to the cloud
where have less control ?
4. Resource management : Not all IT professionals trust the cloud yet. If team was
used to managing physical servers, they might need educating on the new
infrastructure or even reconfiguring to introduce new roles.

(1.4 | Seven-step Model of Migration into a Cloud


¢ Seven-step model is as follows :
1. Different variances of cloud services
Cloud as a tool
wn

Cloud compatible
Current cost
Nowe

To manage rather than operate


To simplify
. Gain more knowledge.
eo Step 1 : Cloud migration assessments comprise assessments to understand the
issues involved in the specific case of migration at the application level or the
code, the design, the architecture, or usage levels.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 1-10 Introduction to Cloud Computing

These assessments are about the cost of migration as well as about the ROI that
can be achieved in the case of production version.
Step 2 : An isolating all systemic and environmental dependencies of the
enterprise application components within the captive data center.
Step 3 : Generating the mapping constructs between what shall possibly remain in
the local captive data center and what goes onto the cloud.
Step 4 : Substantial part of the enterprise application needs to be rearchitected,
redesigned, and reimplemented on the cloud.
Step 5 : We leverage the intrinsic features of the cloud computing service to
augment our enterprise application in its own small ways.
Step 6 : We validate and test the new form of the enterprise application with an
extensive test suite that comprises testing the components of the enterprise
application on the cloud as well.
Step 7 : Test results could be positive or mixed. In the latter case, we iterate and
optimize as appropriate. After several such optimizing iterations, the migration is
deemed successful.

[1.5 | Trends in Computing

1. IT maturity analysis
The technology maturity analysis, which compares the technologies against each
other in terms of their adoption rates and current investment rates.
Major trends that emerged in cloud computing technology are small and medium
business, hybrid cloud model etc.
2. Technology Trends to Watch
Virtualization is technology that lets user create useful IT services using resources
that are traditionally bound to hardware.
Virtualization is composed of server, storage, desktop, application, infrastructure
and network. It is compatible with today's technology.
Data growth : Day by day, data may goes on increasing and organization can not
provide storage. So cloud is solution for storing and processing of growing data.
Green IT : The term green IT means nothing other than the use of information
technology in a manner that preserves resources and protects the environment.
Cloud computing, the provision of IT infrastructure via the Internet, plays an
important role in green IT. Cloud allows the replacement of poorly utilized data
centers with highly optimized server environments. Besides reducing the energy
costs, it also minimizes the time required to maintain a company's proprietary
systems.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 1-11 Introduction to Cloud Computing

1.6 | Cloud Service Models SPPU : March-19


Service models describe the type of service that the service provider is offering.
The best-known service models are Software as a Service, Platform as a Service,
and Infrastructure as a Service.

The service models build on one another and define what a vendor must manage
and what the client's responsibility is.
Service models : This consists of the particular types of services that you can
access on a cloud computing platform.
Cloud service is any service made available to users on demand via the Internet
from a cloud computing provider's servers as opposed to being provided from a
company's OWI on-premises servers.
Cloud services are designed to provide easy, scalable access to applications,
resources and services, and are fully managed by a cloud services provider.
A cloud service can exist as a simple web-based software program with a technical
interface invoked via the use of a messaging protocol, or as a remote access point
for administrative tools or larger environments and other IT resources.
The organization that
provides cloud-based IT
Cloud Clients
resources is the cloud
Web browser, mobile app, thin client, terminal
provider. Cloud providers emulator, ...
normally own the IT
write some points of saas, paas
resources for lease by cloud | and Iaas also
consumers, and could also .
resell IT resources leased 2 SaaS
from other providers. 2 CRM, Email, virtual desktop, communication,
. 3 games, ...
Cloud computing, often
. £
described as a stack, has a = PaaS
broad range of services built ES Execution runtime, database, web server,
a development tools, ...
on top of one another under
the name cloud. laaS
structure
Infra

Virtual machines, servers, storage, load


Fig. 1.6.1 shows cloud balancers, network, ...
computing stack.
Flavors of Cloud computing Fig. 1.6.1 Cloud computing stack
is as follows;
1. SaaS applications are designed for end-users, delivered over the web
2. PaaS is the set of tools and services designed to make coding and deploying
those applications quick and efficient

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge

the above fig is the cloud computing stack which showing the service models of cloud computing
Cloud Computing 1-12 Introduction to Cloud Computing

3. IaaS is the hardware and software that powers it all - servers, storage, networks,
operating systems.
Users can access the application on the internet without installing
the application on local storage thus because of it saves the
EEX] software as a Service (SaaS) computer hardware also.

Model in which an application is hosted as a service to customers who access it


via the Internet.
The provider does all the patching and upgrades as well as keeping the
infrastructure running.
The traditional model of software distribution, in which software is purchased for
and installed on personal computers, is referred to as product.
In this model, the user, client or consumer runs an application from a cloud
infrastructure. Through an interface such as a web browser, the client or user may
access this application from a variety of devices.
The complete application is offered as on demand service. This saves the client
from having to invest in any software licenses or servers up front, and can save
the provider money since they are maintaining and providing only a single
application.
In this model, the client does not manage cloud infrastructure, networks or
servers, storage, or operating systems. Even, Microsoft, Google, and Zoho offer
SaaS.
The SaaS concept can be defined as providing robust "web-based, on-demand
software, storage and various applications” to organizations.
The SaaS model has emerged as an alternative to traditional one-time licensing for
providing and maintaining the software needed by knowledge workers within
organizations. Multiple clients can access the application which is host on Internet
by service provider.
Fig. 1.6.2 shows SaaS. Thus client is able to access the application from anywhere or from
any device with the Internet connection.
Device Comtability issue also does not get occur because of this.
It also becomes easy to managed the application for Service
Client Provider

| Application
: ~f—
Client
H

i
Service provider
offering SaaS
Client

Fig 1.6.2 SaaS

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 1-13 Introduction to Cloud Computing

Characteristics of Saa$ :

1. Software applications or services are stored remotely.


2. A user can then access these services or software applications via the Internet.
3. In most cases, a user does not have to install anything onto their host machine, all
they require is a web browser to access these services and in some cases, a
browser may require additional plug-in/add-on for certain services.
Network-based management and access to commercially available software from
central locations rather than at each customer's site, enabling customers to access
applications remotely via the Internet.
Application delivery from a one-to-many model, as opposed to a traditional
one-to- one model.

Benefits of Saa$ :

1. You only pay for what you use


Easier administration and invoicing
Automatic updates and patch management
Compatibility : All users have access to the same version of software
A

Easier collaboration
It support automated update and patch management services.

EEX] Platform as a Service (PaaS)


Platform as a service is another application delivery model and also known as
cloud-ware. Supplies all the resources required to build applications and services
completely from the Internet, without having to download or install software.
Services include : Application design, development, testing, deployment, and
hosting, team collaboration, web service integration, database integration, security,
scalability, storage, state management, and versioning.
PaaS is closely related to SaaS but delivers a platform from which to work rather
than an application to work with. paas platform deto to build the applications
This model involves software encapsulated and offered as a service, from which
higher levels of service may then be built. The user, customer, or client in this
model is the one building applications which then run on the provider's
infrastructure.
This in turn provides customers and clients with the capability to deploy
applications onto the cloud infrastructure using programming tools and languages,
which the provider supports.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 1-14 Introduction to Cloud Computing

¢ The customer still does not manage the framework, network, servers or operating
system, but has control over deployed applications and sometimes over the
hosting environment itself.
Some examples of Platform as a Service include Google's App Engine or Force.com
PaaS consists of following components :
1. Browser based development studio
2. Pay contrary to billing
3. Management and supervising tools
4. Seamless deployment to host run time environment.
Characteristics of Paa$ :
1. It support multi-tenant architecture.
2. It support for development of group collaboration.
3. PaaS systems can be deployed as public cloud services or as private cloud
services.
Provision of runtime environments. Typically each runtime environment
supports either one or a small set of programming languages and frameworks
Support for custom applications. Support for the development, deployment and
operation of custom applications.
Preconfigured capabilities. Many PaaS systems are characterized by capabilities
that are preconfigured by the provider, with a minimum of configuration
available to developers and customer operations staff.
Support for porting existing applications. While many PaaS systems are
primarily designed to support "born on the cloud” applications.
Security is an important characteristic in PaaS. It needs to provide
authentication and authorization to differentiate the access rights of different
users
Benefits of Paas :
1. Scalability including rapid allocation and deallocation of resources with a
pay-as-you-use model
Reduced capital expenditure
WN

Reduced lead times with on-demand availability of resources


Self-service with reduced administration costs
Ue

Reduced skill requirements


Support of team collaboration
No

Ability to add new users quickly.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 1-15 Introduction to Cloud Computing

1.6.3 | Infrastructure as a Service (laaS)

IaaS gives the storage room likeness to the in-house datacenter stood out from
various organizations sorts.
Center datacenter framework segments are capacity, servers (registering units), the
system itself, and administration apparatuses for foundation upkeep and checking.
Each of these parts has made a different market specialty. While some little
organizations have practical experience in just a single of these IaaS cloud
specialties, vast cloud suppliers like Amazon or Right Scale have offerings over all
Taa$S territories.
Fig. 1.6.3 shows IaaS.

Physical server

Enterprise
Data storage
device
Computer

User

Fig. 1.6.3 laaS

It offers the hardware so that your organization can put whatever they want onto
it. Rather than purchase servers, software, racks, and having to pay for the
datacenter space for them, the service provider rents those resources :
1. Server space 2. Network equipment
3. Memory 4. CPU cycles 5. Storage space
Again, the customer is not managing cloud infrastructure, but in this case, the
customer does control operating systems, deployed applications, storage, and
sometimes-certain networking components
Examples : Amazon EC2, Rackspace Mosso, GoGrid

TaaS server types :


1. Physical server : Actual hardware is allocated for the customer's dedicated use.
2. Dedicated virtual server : The customer is allocated a virtual server, which
runs on a physical server that may or may not have other virtual servers.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 1-16 Introduction to Cloud Computing

3. Shared virtual server : The customer can access a virtual server on a device
that may be shared with other customers.

Advantages of laa$S :
1. Elimination of an expensive and staff-intensive data center
Ease of hardware scalability
Nn

Reduced hardware cost


Noe»

On-demand, pay as you go scalability


Reduction of IT staff

Suitability for ad hoc test environments


Allows complete system administration and management
Support multiple tenants

EEX] Difference between laaS, PaaS and SaaS


IaaS PaaS SaaS

TaaS gives users automated PaaS provides a framework SaaS makes applications
and scalable environments for quickly developing and available through the internet.
deploying applications

Amazon Web Services, for Google Cloud Platform SaaS applications such as
example, offers IaaS through provides another PaaS option Gmail, Dropbox, Salesforce, or
the Elastic Compute Cloud, or in App Engine Netflix
EC2

In Iaas, infrastructure as a In Paas, platform as a service In SaaS, software as a service


service.

Virtual platform on which Operating environment was Operating environment largely


required operating included. irrelevant, fully functional
environment and application application provided
deployed
IaaS is a cloud service that PaaS refers to cloud platforms SaaS allows people to use
provides basic computing that provide runtime cloud-based web applications.
infrastructure: servers, storage, environments for developing,
and networking resources. In testing, and managing
other words, Iaa$ is a virtual applications
data center

Major IaaS providers include Examples of PaaS services are email services such as Gmail
Amazon Web Services, Heroku and Google App and Hotmail are examples of
Microsoft Azure, and Google Engine. cloud-based SaaS services.
Compute Engine.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 1-17 Introduction to Cloud Computing

IaaS services are available on ~~ PaaS solutions are available SaaS services are usually
a pay-for-what-you-use model with a pay-as-you-go pricing available with a pay-as-you-go
model. pricing model

Used by IT administrator Used by software developers Used by end user

1. Explain benefits of SaaS in detail. SPPU : March-19, In Sem, Marks 4

Cloud Computing Logical Architecture

e Fig. 1.7.1 shows architectural framework of cloud computing.


Users /
Brokers Pel = & 8

7] I

x 7 [4 \r __r x
|
Service request examiner and I
admission control !
- Customer - driven service management |
- Computation risk management !
- Automatic resource management

SLA i
Resource !
Allocator = - \
Pricing Accounting 1
||
i
VM Service request |
Monitor Dispatcher monitor !
|
|
|
|

i|
Virtual !
Machines I
(VMs)
|
I
|
|
|

i|
Physical !
Machines

i
I
|
|

Fig. 1.7.1 Architectural framework

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 1-18 Introduction to Cloud Computing

1. Users/Brokers : They submit their service requests from anywhere in the world
to the cloud.
2. SLA resource allocator : It is a kind of interface between users and cloud
service provider which enable the SLA-oriented resource management.
3. Service request examiner and admission control : It interprets the submitted
request for QoS requirements before determining whether to accept or reject
the request. Based on resource availability in the cloud and other parameters
decide.
4, Pricing : It is in charge of billing based on the resource utilization and some
factors. Some factors are request time, type etc.
5. Accounting : Maintains the actual usage of resources by request so that the
final cost can be charged to the users.
6. VM monitor : Keeps tracks on the availability of VMs and their resources.
7. Dispatcher : The dispatcher mechanism starts the execution of admitted
requests on allocated VMs.
8. Service request monitor : The request monitor mechanism keeps track on
execution of request in order to be in tune with SLA.

Cloud computing service layers :

Parameters Services Description

Application Focused Services Services - Complete business services such as PayPal,


OpenID, OAuth, Google Maps, Alexa
Application Application - Cloud based software that eliminates the
need for local installation such as Google Apps, Microsoft
Online
Development Development - Software development platforms used to
build custom cloud based applications (PAAS and SAAS)
such as SalesForce

Infrastructure Platform Platform - Cloud based platforms, typically provided


Focused using virtualization, such as Amazon ECC, Sun Grid

Storing Storage - Data storage or cloud based NAS such as


CTERA, iDisk, CloudNAS

Hosting Hosting - Physical data centers such as those run by IBM,


HP, NaviSite, etc.

Cloud components :
¢ Cloud computing solutions are made up of several elements.
1. Clients : Mobile, terminals or regular computers.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 1-19 Introduction to Cloud Computing

2. Benefits : Lower hardware costs, lower IT costs, security, data security, less
power consumption, ease of repair or replacement, less noise.
3. Data centers : Collection of servers where the application to subscribe is
housed. It could be a large room in the basement of your building or a room
full of servers on the other side of the world
4. Virtualizing servers : Software can be installed allowing multiple instances of
virtual servers to be used and a dozen virtual servers can run on one physical
server.
5. Distributed servers : Servers don't all have to be housed in the same location.
It can be in geographically disparate locations. If something were to happen at
one site, causing a failure, the service would still be accessed through another
site. If the cloud needs more hardware, they can add them at another site.

Cloud Computing Service Architecture


¢ Cloud computing services are divided into three classes :
1) Infrastructure as a service,
2) Platform as a service
3) Software as a service.
e Fig. 1.7.2 shows the layered organization of the cloud stack from physical
infrastructure to applications.

Cloud applications Web browser Software as a service

Cloud
Cloud platform development Platform as a service
environment

Virtual Infrastructure as a
Cloud infrastructure infrastructure service
manager

Service content Main access Service class


and management tool

Fig. 1.7.2 Cloud stack

e Cloud services are designed to provide easy, scalable access to applications,


resources and services and are fully managed by a cloud services provider.
e A cloud service can exist as a simple web-based software program with a technical
interface invoked via the use of a messaging protocol or as a remote access point
for administrative tools or larger environments and other IT resources.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 1-20 Introduction to Cloud Computing

¢ The organization that provides cloud-based IT resources is the cloud provider.


Cloud providers normally own the IT resources for lease by cloud consumers, and
could also resell IT resources leased from other providers.
e Cloud computing, often described as a stack, has a broad range of services built
on top of one another under the name cloud.
1. Software as a service
e SaaS utilizes the internet to deliver applications, which are managed by a
third-party vendor, to its users. A majority of SaaS applications run directly
through your web browser, which means they do not require any downloads or
installations on the client side.
e Salesforce.com uses SaaS model. SaaS applications are designed for end-users,
delivered over the web.
2. Infrastructure as a service
e JaaS is a cloud computing service where enterprises rent or lease servers for
compute and storage in the cloud. Users can run any operating system or
applications on the rented servers without the maintenance and operating costs of
those servers.
e Amazon Web Services mainly offers IaaS.
e IaaS is the hardware and software that powers it all - servers, storage, networks,
operating systems.
3. Platform as a service
e Platform as a Service (PaaS), provide cloud components to certain software while
being used mainly for applications. PaaS delivers a framework for developers that
they can build upon and use to create customized applications.
e All servers, storage, and networking can be managed by the enterprise or a
third-party provider while the developers can maintain management of the
applications.
¢ Google AppEngine, an example of Platform as a Service.
e PaaS is the set of tools and services designed to make coding and deploying those
applications quick and efficient.

EE] Developing Holistic Cloud Computing Reference Model


e Cloud computing is an internet-based computing where software, shared resources
and information are served to devices such as computers, electricity grid. In these
computing techniques, how the work is getting done is hidden from its users. It
describes a new consumption, delivery and supplement model for IT services.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 1-21 Introduction to Cloud Computing

¢ Cloud computing is one of the next significant stage in the Internet's evolution,
providing the means through which everything from computing power to
computing infrastructure, applications, business processes to personal collaboration
can be delivered to the user as a service wherever and whenever they need.
e The "cloud" in cloud computing can be defined as the set of hardware, networks,
storage, services and interfaces that combine to deliver aspects of computing as a
service.
e Consumer cloud computing services has been well established ever since
mainstream Internet. Known examples are Webmail services and social networking
platforms.
e However the adoption of cloud computing within the enterprise sector has been
slow. This slow uptake in cloud services that promises so much has been
primarily influenced by the numerous security risks, concerns and challenges
posed within such an environment.
¢ Governance, risk and compliance factors of cloud services need to be fully
assessed and evaluated by organizations to provide informed judgments. Data and
information lifecycle, source and origination, transfer, destination, validation and
deletion all need to be understood.

1.8.1 | Complete View of Cloud Management

e Despite significant attention and substantial efforts both in industry and academia,
cloud computing has not yet reached its full potential. Commonly stated obstacles
for cloud adoption include confusion about terminology with multiple delivery
models (SaaS, PaaS, IaaS) and deployment cloud scenarios.

¢ The analysis of cloud computing obstacles in five high-level research challenges :


1. Service life cycle optimization for improved construction, deployment and
operation of cloud services.
2. Adaptive self-preservation with increased autonomy in resource management,
enabling fewer administrators to handle increasingly larger systems.
3. Cloud and service self-management based on non-functional management
criteria, extending beyond typical cost-performance trade-offs to also
incorporate aspects of trust, risk and eco-efficiency in decision making.
4. Support for service deployment and operation in multiple cloud architectures
such as private clouds, cloud federation, cross-cloud service provisioning.
5. Market and legislative issues including identification of new market roles and
business models for clouds.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 1-22 Introduction to Cloud Computing

1.8.2 | Cloud Computing Reference Model

Cloud Computing Reference Model (CC-RM) is developed to facilitate the process


of cloud modeling, deployment planning and architecture.
The cloud computing reference model is an abstract model that characterizes and
standardizes the functions of a cloud computing environment by partitioning it
into abstraction layers and cross-layer functions.
The cloud computing reference model will establish a cloud modeling and
architecture foundation from which an organization can realistically plan, model,
architect and deploy cloud computing in a pragmatic fashion to address real and
pressing business and technical challenges.
Cloud should not be treated as a solution looking for a problem but as a collection
of cloud patterns that can be configured to meet a wide array of business and
technical requirements.
Elements of cloud computing reference model is as follows :
a. Cloud enablement model
b. Cloud deployment model
c. Cloud governance and operations model
d. Cloud ecosystem model.
The cloud enablement model describes the tiers of cloud computing foundation,
enablement and business capabilities provided by cloud platform and service
providers to potential consumers of cloud-enabled technology and business
capabilities.
The Cloud Deployment Model (CDM) describes the range of cloud deployment
scenarios available to the enterprise internal-private cloud, external public cloud,
hybrid cloud and community clouds.
Cloud governance and operations model : It describes the governance, security
and privacy, operations and support, management and monitoring requirements
for cloud computing and ensures that the user has considered all the potential
operational risks for adopting cloud for their enterprises.
The cloud ecosystem model considers the requirements of developing and
sustaining a cloud ecosystem comprised of cloud providers, cloud consumers,
cloud intermediaries, as well as the cloud network.
Fig. 1.8.1 shows cloud computing reference model.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 1-23 Introduction to Cloud Computing

Interfaces for web 2.0

Software as a Service

e Scientific applications, End user applications


User 1 — e Social networking
e Examples : facebook, Google documents, Flickr

user2 | &&——p»
Platform as a Service

® Runtime environment for applications


User 3 < — >
e Development and data processing platforms
® Examples : Windows Azure, Hadoop, Aneka

UserN | {>
Infrastructure as a Service

eo Virtualized servers
e Storage and networking
® Examples : Amazon EC2, S3, vCloud

Fig. 1.8.1 Cloud computing reference model

Software as a Service
e SaaS provider dispose the applied software unified on their server, the user can
subscribe applied software service from the manufacturer through Internet.
e Provider supply software pattern through Browser, and charge according to the
quantity of software and using time. The advantage of this kind of service pattern
is that the provider maintains and manages software, supplies the hardware
facilities, the users can use software everywhere when they own the terminal
which can log in Internet.
Platform as a Service (PaaS) :
e PaaS takes develop environment as a service to supply. This layer provides a
platform for creating applications.
e PaaS solutions are essentially development platforms for which the development
tool itself is hosted in the cloud and accessed through a browser.
e With PaaS, developers can build web applications without installing any tools on
their computers and then deploy those applications without any specialized
systems administration skills.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 1-24 Introduction to Cloud Computing

¢ Google App Engine is the representative product through their server and
Internet.

Infrastructure as a Service (laas) :


e In this layer, servers, network devices, and storage disks are made available to
organizations as services on a need-to basis.
¢ IaaS takes infrastructure which is made of many servers as a measurement service
to the customers.
e It integrates memory and I/O devices, storage and computing ability into a virtual
resources pool, and provides storage resources and virtualization service for the
whole industry.
e This is a way of hosted hardware, and the customer pays when they use the
hardware.
¢ For example, Amazon Web Service and IBM Blue Cloud all rent the infrastructure
as a service.

(1.9 | Cloud System Architecture


e There are several factors that user need to take into consideration before designing
own cloud-based systems architecture, particularly if user is considering a
multi-cloud/region architecture.
1. Speed - The cloud gives more flexibility to control the speed or latency of
site/application.
2. Complexity - Before constructing highly customized hybrid cloud solution
architecture, make sure user properly understand the actual requirements of
application, SLA, etc

3. Cloud portability - Although it might be easier to use one of the cloud


provider's tools or services, such as a load balancing or database service.
4. Cost - Before launching servers, user should clearly understand the SLA and
pricing models associated with cloud infrastructure(s)
5. Security - Check security measurement provided by vendor.

Single Server
e A single server means only one server, whether physical or virtual, containing one
web server, database, and application.

eo LAMP stack is an example of single server. This structure is not widely utilized
due to associated security risks.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 1-25 Introduction to Cloud Computing

(1.9.1 | Single Cloud Site Architectures

Fig. 19.1 shows single cloud site


architecture. DNS
Cloud contains load balancer server,
application server and database. I
I
I

A single cloud environment is


I
I Load balancer server
I
accomplished by using a single cloud I
I

provider to serve any and all


Application
applications or services that the !
organization decides to migrate to the
cloud. Single cloud environments can I
I

utilize either private or public clouds, I


I Database
using whichever one better serves
their current and future needs. CER NEE
Organizations with a single cloud
model are employing the cloud for a
Database (Cloud storage)
single service or application, such as
email, Enterprise Resource Planning
(ERP), Customer Relationship Fig. 1.9.1 Single cloud site architecture
Management (CRM), or similar.
A single cloud environment may be more fitting for smaller or less technically
adept organizations that would like to gain the many benefits of the cloud without
it becoming overwhelming.

(1.9.2 | Redundant 3-Tier Architecture

Fig. 1.9.2 shows 3-tier architecture.


Redundant three-tier architectures add another set of the same components for
redundancy. Additional design components do increase complexity, but are
required if designing for failover and recovery protection.
Designing redundant infrastructures requires a well thought out plan for the
components within each layer (horizontal scaling), as well as a plan for how the
traffic will flow from one layer to another (vertical scaling).
Horizontal scaling means that user can scale by adding more machines into pool
of resources. For example, as traffic increases, add more web servers to take on
the traffic.
Vertical scaling means that, scale by adding more power ( CPU, Memory) to an
existing machine.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 1-26 Introduction to Cloud Computing

rd AN
Load balancer 1 Load balancer 2

Application 1 Application 2

Snapshot
backups

Fig. 1.9.2 3-tier architecture

e A failsafe system comprised of redundant cloud storage devices is established so


that when the primary device fails, the redundant secondary device takes its place.

EEX] Multi-Data Centre Architecture


¢ Fig. 1.9.3 shows multi-data centre architecture.
eo If cloud infrastructure supports multiple data centers, it's recommended that
organization spread system architecture across multiple data centers to add
another layer of redundancy and protection.
e Each data center in a cloud is designed to be an isolated segment inside the same
geographical cloud. So if a power failure occurs in one data center, the other data
centers will be unaffected.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 1-27 Introduction to Cloud Computing

DNS

Datacenter 1 / > Datacenter 2

Load balancer 1 Load balancer 2

Application 1 Application 2

= amd
Replication Slave <a»

Snapshot
backups

Fig. 1.9.3 Multi-data centre architecture

EET] cloud Deployment Models from notes pdf

e Cloud deployment models are refers to the location and management of the
cloud’s infrastructure.
¢ Deployment models are defined by the ownership and control of architectural
design and the degree of available customization. Cloud deployment models are
private public and community clouds.
e Fig. 1.10.1 shows cloud deployment model.
1. Public cloud :
¢ The cloud infrastructure is made available to the general public or a large industry
group and is owned by an organization selling cloud services.
e Public cloud is a huge data centre that offers the same services to all its users. The
services are accessible for everyone and much used for the consumer segment.
e Examples of public services are Facebook, Google and LinkedIn.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 1-28 Introduction to Cloud Computing

Private cloud

/_aa8
“2
(a) Private cloud (b) Public cloud

Fig. 1.10.1
e Public cloud benefits :
a) Low investment hurdle : Pay for what user use.
b) Good test/development environment for applications that scale to many
servers.
Public cloud risks :
a) Security concerns : Multi-tenancy and transfers over the Internet.
b) IT organization may react negatively to loss of control over data center
function.

2. Private cloud :
The cloud infrastructure is operated solely for a single organization. It may be
managed by the organization or a third party and may exist on-premises or
off-premises.
Private cloud benefits :
a) Fewer security concerns as existing data center security stays in place.
b) IT organization retains control over data center.
Private cloud risks :
a) High investment hurdle in private cloud implementation, along with purchases
of new hardware and software.
b) New operational processes are required; old processes not all suitable for
private cloud.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 1-29 Introduction to Cloud Computing

Organization 1 Organization 2

Fig. 1.10.2 Community cloud

3. Community cloud :
e The cloud infrastructure is shared by several organizations and supports a specific
community that has shared concerns (e.g. mission, security requirements, policy, or
compliance considerations). It may be managed by the organizations or a third
party and may exist on-premises or off-premises.
4. Hybrid cloud :
eo The cloud infrastructure is a composition of two or more clouds (private,
community or public) that remain unique entities but are bound together by
standardized or proprietary technology that enables data and application
portability (e.g., cloud bursting for load-balancing between clouds).
e Hybrid cloud benefits :
a) Operational flexibility : Run mission critical on private cloud, dev/test on
public cloud
b) Scalability : Run peak and bursty workloads on the public cloud
¢ Hybrid cloud risks :
a) Hybrid clouds are still being developed; not many in real use

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 1-30 Introduction to Cloud Computing

b) Control of security between private and public clouds, some of same concerns
as in public cloud

EETX] pifference between Public and Private Cloud


Public cloud Private cloud

Public cloud infrastructure is offered via web Private cloud infrastructure is dedicated to a
applications and also as web services over single organization.
Internet to the public.

Support multiple customer Support dedicated customer


Full utilized of infrastructure. Does not utilize shared infrastructure

Security is low as compared to private cloud High level of security


Low cost High cost

Azure, Amazon Web Services, Google App An example of the Private Cloud is NIRIX’s
Engine and Force.com are a few examples of one Server with dedicated servers
public clouds

1. Explain cloud deployment models as per the NIST guidelines.


SPPU : March-19, In Sem, Marks 6

EKER Multiple Choice Questions


Q.1 Point out the wrong statement :

a| Abstraction enables the key benefit of cloud computing : shared, ubiquitous


access.
b| Virtualization assigns a logical name for a physical resource and then provides a
pointer to that physical resource when a request is made.
c| All cloud computing applications combine their resources into pools that can be
assigned on demand to users.
d| All of the mentioned.

Q.2 Point out the wrong statement :

a| The massive scale of cloud computing systems was enabled by the


popularization of the Internet.
b| Soft computing represents a real paradigm shift in the way in which systems are
deployed.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 1-31 Introduction to Cloud Computing

Cloud computing makes the long-held dream of utility computing possible with
a pay-as-you-go, infinitely scalable, universally available system.
ld All of the mentioned.

Q3 Which of the following is essential concept related to Cloud ?

[a] Reliability [b] Productivity


Abstraction [d] All of the mentioned.

Q4 Point out the wrong statement :

[al All applications benefit from deployment in the cloud.


bl With cloud computing, you can start very small and become big very fast.
Cloud computing is revolutionary, even if the technology it is built on is
evolutionary.
ld None of the mentioned.

Q5 Which of the following cloud concept is related to pooling and sharing of


resources ?

[a] Polymorphism [b] Abstraction


Virtualization [d] None of the mentioned

Q.6 CDC Stands for .


[a] Cloud Data Computing [b] Cloud Data Cluster
Cloud Data Center [d] Computing Data on Cloud

Q.7 In which year IoT was introduced ?

[a] 1999 [b] 1998


1996 [d] 1997
Q.8 is mainly used to utilize idle resources in the nodes.

[a] Dedicated cluster [b] Enterprise cluster


Distributed cluster [d] Centralized cluster

Q.9 is the entry point into the cloud for user and administrators.
[a] Cloud Manager [b] Group Manager
Instance Manager [d] VM Manager

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 1-32 Introduction to Cloud Computing

Q.10 The services that provides utility may directly correlate with their ?

[al expectations [b] profit


time [d] satisfaction

Q.11 The promise of has raised the IT expectations of small and medium
enterprises beyond measures.

[a] cloud computing [b] cloud computing


cluster computing [d] client-server computing

Q.12 Cloud supports applications and resources elasticity.

[a] performance [b] reliability


scalability [d| none of the above

Q.13 cloud is shared by several organizations and supports a specific community


that has shared concerns.

[a] Private [b] Public


Managed [d| Community

Q.14 of grids/clouds to offer standard interfaces for dynamically scalable services


delivery in their products.

[a] Vendors [b] Consumers


Integrators [d| Providers

Q.15 VAN Stands for .

[a] Virtual Application Network [b] Virtual Acceptable Network


Virtual Admission Network [d] Virtual Area Network

Q.16 In what type of cluster, the nodes are closely packaged in one or more racks sitting
in a room, and the nodes are not attached to peripherals ?

[a] Compact [b] Slack


loosely coupled [d] Tightly coupled

Q.17 Virtualization is a computer architecture technology by which multiple _____ are


multiplexed in the same hardware machine.

[a] virtual memory [b] virtual machines


physical machines [d] virtual machine monitor

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 1-33 Introduction to Cloud Computing

Q.18 Which of the following is not VI manager ?


[al Apache VCL [b] App Logic
Google VI [d| Nimbuz 3
Q.19 According to Infosys how many steps are there in the migration model ?

[a] 4 [b] 5
[c] 6 ld] 7
Q.20 _____ is the process of transferring data between storage types, formats or systems.

[a] Data mediation [b] Data integrity


Data modification [d| Data migration

Q.21 Which of the following is not principle of the cloud ?

[a] Federation [b] Non independent


Trust [d] Isolation

Q22 _ is a directory on the cluster node where a VM is running.


[a] Virtualization [b] KVM
Virtual Machine Directory [d] VMware

Answer keys for Multiple Choice Questions :

Q1 c Q.2 b Q3 c Q4 a

Q5 c Q.6 c Q7 a Qs b

Q.9 a Q.10 a Q.11 b Q.12 c

0.13 d Q.14 a Q.15 d Q.16 a


017 b Q.18 b Q.19 d 0.20 d
0.21 a 022 b

aaa

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 1-34 Introduction to Cloud Computing

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


UNIT - 11

Data Storage and Cloud Computing

Syllabus
Data Storage : Introduction to Enterprise Data Storage, Direct Attached Storage, Storage Area
Network, Network Attached Storage, Data Storage Management, File System, Cloud Data Stores,
Using Grids for Data Storage. Cloud Storage : Data Management, Provisioning Cloud storage,
Data Intensive Technologies for Cloud Computing. Cloud Storage from LANs to WANs : Cloud
Characteristics, Distributed Data Storage

Contents
2.1 Introduction to Enterprise Data Storage
2.2 Data Storage Management
2.3 File System LL... March-19,20, ----------- Marks 6
2.4 Cloud Data Stores ~~ .................. Dec.-16,17, March-19, - - - - Marks 5
2.5 Using Grids for Data Storage
2.6 Cloud Storage
2.7 Cloud Storage from LANs to WANs
2.8 Multiple Choice Questions

Enterprise Data Storage:-


Enterprise storage is a centralized repository for business information that provides
common data management, protection and sharing functions through connections to
computer systems.

Because enterprises deal with heavy workloads of business-critical information, enterprise


storage systems should be scalable for workloads of hundreds of terabytes or even
petabytes without relying on excessive cabling or the creation of subsystems. Other
important aspects of an enterprise storage system are unlimited connectivity and support
for multiple platforms.
(2-1)
Cloud Computing 2-2 Data Storage and Cloud Computing

(2.1 | Introduction to Enterprise Data Storage


Enterprise storage is a centralized repository for business - critical information that
provides data sharing, data management and data protection across multiple
computer systems.
Enterprise storage may include a Storage Area Network (SAN), Network -
Attached Storage (NAS) devices and Direct Attached Storage (DAS).

(2.1.1 | Direct Attached Storage

Direct - Attached Storage (DAS) is hard disk drives or solid - state drives
connected directly inside or outside (in a storage enclosure) to a single computer
or server that cannot be accessed by other computers or servers. DAS is not
networked through Ethernet or FC switches.
The storage device may include one or more drives built into a server and with an
appropriate host bus adapter, may be configured as a RAID array.
Examples of DAS include hard drives, optical disc drives and storage on external
drives.
One advantage of DAS storage is its low initial cost. A key disadvantage of DAS
storage is its limited scalability. A Host Bus Adaptor can only support a limited
number of drives.
Direct attached storage tends to be less expensive than NAS and SAN
implementations, but lacks scalability and flexibility.
DAS are of two types : Internal DAS and external DAS.
1. Internal DAS : Storage device is internally connected to the host. The
connection between host and storage device is serial or parallel bus.
2, External DAS : Server connects directly to the external storage device.

Advantages of DAS :
1. DAS is ideal for localized file sharing in environments with a single server or a
few servers.
DAS devices can offer block-level access or file-level access.
DAS also offers ease of management and administration.
ISA
CR

DAS can still be used locally to store less critical data.


Initial cost of DAS is lower than NAS.
Disadvantages of DAS :
1. DAS is limited in its scalability.

2. DAS is limited to dedicated servers.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 2-3 Data Storage and Cloud Computing

3. Unused resources cannot be reallocated easily.


4. If the server the device is attached to is down for any reason, the data stored
on attached DAS devices is inaccessible.

2.1.2 | Storage Area Network

Storage Area Network (SAN) is a dedicated high-performance network or


subnetwork dedicated to storage that is independent of an organization's common
user network.
It interconnects pools of disk or solid - state storage and shares it to multiple
servers so each one can access data as if it was directly attached. A storage area
network is a dedicated network that provides access to consolidated, block level
data storage.
SANs are primarily used to enhance storage devices, such as disk arrays, tape
libraries and accessible to servers so that the devices appear like locally attached
devices to the operating system.
SAN systems provide a simple, untyped, fixed-size (block), memory-like interface
for manipulating non-volatile magnetic media. SAN is used when multiple servers
need access to the same data (Example : Databases).
A SAN is a high - speed network of storage devices, separate from the local area
network that also connects those storage devices with servers. It provides block -
level storage that can be accessed by the applications running on any and all
servers connected to the SAN.
SAN storage devices can include tape libraries and more commonly, disk-based
devices, like RAID hardware.

Fig. 2.1.1 shows storage area network.

Client ] LAN / WAN a]

Server

Storage
area
network

Disks BE TE

Fig. 2.1.1 Storage area network

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 2-4 Data Storage and Cloud Computing

e SANs are primarily used to enhance storage devices, such as disk arrays, tape
libraries and accessible to servers so that the devices appear like locally attached
devices to the operating system
e SAN systems provide a simple, untyped, fixed-size (block), memory-like interface
for manipulating nonvolatile magnetic media. SAN is used when multiple servers
need access to the same data (Example : Databases).
e A SAN is a high - speed network of storage devices, separate from the local area
network that also connects those storage devices with servers. It provides block -
level storage that can be accessed by the applications running on any and all
servers connected to the SAN. SAN storage devices can include tape libraries and
more commonly, disk - based devices, like RAID hardware.
e A storage area network is designed for managing very large amounts of network
storage. For large organizations with many servers that need access to the same
data, SAN offers better performance and flexibility than DAS or NAS.
e Storage area network is defined as a set of interconnected devices and servers that
are connected to a common communication and data transfer infrastructure such
as fiber channel. The purpose of the SAN is to allow multiple servers access to a
pool of storage in which any server can potentially access any storage unit.
e A SAN is a network designed to transfer data from servers to targets and it is
alternative to directly attached target architecture or to a DAS architecture, where
the storage is connected to the servers on general purpose networks.
eo A SAN can be considered as an extended and shared storage bus within a data
center, consisting of various storage devices and specific interfaces (e.g. fibre
channel) rather than the Ethernet. SANs provide high - bandwidth block storage
access over long distance via extended fiber channel links.
SAN features :
1. Network resource used exclusively for the storage.
2. SCSI protocol for end to end communication : Minimum impact for DAS to
SAN migration.
Access to data through logical blocks and not to file.
Uk Ww

Compatibility with an high number of nodes.


High reliability and ability to react to failures.
Performance allows concurrent access of disk or tape arrays by two or more
oN

servers at high speeds.


7. Availability have disaster tolerance built in, because data can be mirrored
using a SAN up to 10 km or 6.2 miles away.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 2-5 Data Storage and Cloud Computing

Scalability like a LAN/WAN, it can use a variety of technologies. This allows easy
relocation of backup data, operations, file migration and data replication between
systems.
Advantages of SAN :
SAN support large and heterogeneous block data transfer
ok whe

It gives data transfer reliability


Noe

Reduces LAN traffic


Configuration flexibility
High performance
High scalability
Centralized management
Multiple vendor offerings

Resilience to failure.

EXE) Network - attached Storage


NAS is a file level computer data storage server connected to a computer network
providing data access to a heterogeneous group of clients. It is specialized for
serving files either by its hardware, software or configuration.
NAS devices are storage arrays or gateways that support file - based storage
protocols such as NFS and CiFS and are typically connected via an IP network.
Components of NAS is NIC, OS, 1. CPU and memory, NFS and CIFS, Storage
protocol
Fig. 2.1.2 shows component of NAS.
(See Fig. 2.1.2 on next page)
NAS uses industry standard storage protocols (ATA, SCSI and FC) for connecting
and managing physical disk resources. Network file system and common internet
file system protocols used for file sharing. Network interface card is used for
providing network connectivity.
NAS devices are the most appropriate for high - demand engineering applications
like computer - aided design and software development.
If you have both UNIX and windows users on your network and you want both
groups to be able to share files, NAS devices are most suitable. NAS devices can
make use of existing directories of user accounts from Windows, Netware or
UNIX server.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 2-6 Data Storage and Cloud Computing

Server 1 Server 2

\ /
/

IP network

— Interface for network


= NFS CIFS
Operating system
Interface for storage

Storage arrays

Fig. 2.1.2 NAS components


NAS File 1/0 :
In NAS, there are two file level protocols which allows the file sharing in a NAS.
1. CIFS : Common internet file System, this protocol allows windows
clients /users to share, store and retrieve files on a NAS device
2. NFS : Network file system, this protocol allows Unix clients/users to share,
store and retrieve files on a NAS device
NAS uses file level access for all types of I/O operations. File level I/O is a high
level type of request that specified only file to access. In this file level storage, the
storage disk is configured with a protocol such as NFS or CIFS and the files are
stored and accessed from it in bulk.
It stores files and folders and the visibility is the same to the clients accessing and
to the system which stores it.
File I/O request specifies the file and offset into the file.
1. File system and remote file sharing : File system is a structured way of storing
and organizing data files.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 2-7 Data Storage and Cloud Computing

2. Accessing file system : File system must be mounted before it can be used.
During boot process, operating system mounts the local file system. To mount
any file system, two pieces of information must be specified :
a. A means of uniquely identifying the desired disk drive and partition, such
as device file name, file system label.

b. A directory under which the mounted file system is to be made available


Mount utility allows a super user to splice the root directory of a file system into
the existing directory hierarchy. File system may be detached from the main
hierarchy using the un-mount utility.
NAS places a network between client and file system. NAS systems contain the
file system that interprets metadata, so they do not send as much metadata across
the network. NAS systems have traditionally trusted less, mainly the operating
systems of their clients but not the users of these clients.
A NAS device does not need to be located within the server but can exist
anywhere in a LAN and can be made up of multiple networked NAS devices.
NAS systems usually contain one or more hard disks, often arranged into logical,
redundant storage containers or RAID arrays.
NAS does not provide any of the activities that a server in a server - centric
system typically provides, such as e-mail, authentication or file management. NAS
allows more hard disk storage space to be added to a network that already utilizes
servers without shutting them down for maintenance and upgrades.
NAS systems contain one or more hard disks, often arranged into logical,
redundant storage containers or RAID arrays.
NAS Implementation :
NAS is implemented in two ways : Integrated and gateway.
1. Integrated NAS :
Integrated NAS contains all the components of NAS so it makes the integrated
NAS a self contained environment. Fig. 2.1.3 shows integrated NAS. Integrated
NAS is useful when customer is looking at an integrated NAS which has its own
storage.

Integrated network Windows operating systems client


attached storage

UNIX system client

Fig. 2.1.3 Integrated NAS

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 2-8 Data Storage and Cloud Computing

¢ An integrated NAS solution is available with low - end device to high end
solution. They are connected to the local area network and communicate with file
systems through Trans - Mission Control Protocol / Internet Protocol (TCP / IP).
2. Gateway NAS :
eo It is NAS device with the internal processor but without integrated storage.
Instead, the NAS device connects to storage by direct attachment or by a SAN.
e In a NAS gateway, the proprietary RAID coniroller is replaced with an open,
ANSI-standard SAN interface, usually a fibre channel host bus adapter. This
allows the NAS gateway to use shared, nonproprietary storage, which provides
the flexibility for an effective and scalable storage strategy.
e When deploying NAS gateways, it is important to select the right shared storage
platform to house your organization's data.
¢ Fig. 2.1.4 shows NSA gateway.

FC
Integrated network switch Fibre
attached storage channel
SAN

UNIX system client Windows operating


systems client

Fig. 2.1.4 Gateway NAS


e A front end to disk storage that is highly scalable. A NAS gateway functions like
a diskless file server with unlimited storage. Unlike the traditional NAS device,
which contains the file sharing protocol for access, the file system for storage and
the disks themselves, NAS gateway contains only the file sharing protocol and the
file system.
e The front end of the NAS gateway connects to the LAN over Ethernet. The back
end connects to a RAID array via fibre Channel or to a fibre Channel switch, in
which case the NAS gateway becomes a front end to a SAN.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 2-9 Data Storage and Cloud Computing

Difference between Integrated NAS and Gateway NAS :

Sr. No. Integrated NAS Gateway NAS

1. Integrated NAS uses fibre channel Gateway NAS uses both fibre channel
disks only. and SATA disks.
2. Integrated NAS manages storage Gateway NAS relies on storage array
devices through NAS software. management software to manage
storage devices.
3. It uses specialized operating systems. It uses generic operating systems.

4. Integrated NAS allows both file and Gateway NAS allows only block level
block level access to the clients. access to the clients.

e Benefits of NAS :
1. NAS support NFS and CIFS file system.
Entry level NAS systems are quite inexpensive.
NAS ensure security and user authentication.
ISA

It helps for easy and fast communication.


NAS appliances already include integrated mechanisms for backup, data
synchronization and data replication.
6. Simple to administrate, even remotely, i.e. from another location.
7. NAS device may support one or more RAID levels to make sure that individual
disk failures do not result in loss of data.
8. NAS is a much simpler and less expensive technology.

En Comparison between NAS and SAN

Sr. No. NAS SAN

1 Machine connected with LAN may Server class devices that are equipped
utilize NFS, CIES or HTTP protocol to with SCSI and fibre channel adapters
connect to a NAS connect to a SAN.

2. File system is managed by the NAS The SAN servers manage the file system
head unit
3. Backups and mirrors are generated on Backups and mirrors require a block by
files, not blocks block copy operation.
4. A NAS identifies the data by file name SAN addresses the data by logical block
and byte offset, transfers file data or numbers and transfers the data in disk
metadata blocks.
5. NAS uses TCP/IP networks SAN uses fibre channel

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 2-10 Data Storage and Cloud Computing

2.1.5 | Comparison between DAS, NAS and SAN

Sr. No. Parameter DAS NAS SAN

1 1/0 Protocol SCSI NFS and CIFS SCSI


2. Data sharing Not supported Supported SAN support data
sharing with the
help of specialized
software
3. Capacity sharing Manual Yes Yes
4. Media FC, parallel SCSI Ethernet Fiber channel
5. Storage Storage is directly NAS device is Storage resides on
attached by a cable attached to a a dedicated
to the processor TCP/IP-based network
network

(2.2 | Data Storage Management


e Large number of databases is generated in private organization and government
organization. It is growing out of control. Huge databases can be difficult to
query, challenging to secure and maintain and cumbersome to replicate for testing
and development purposes.

(2.2.1 | Data Storage Management Tools

¢ Storage monitoring software helps user to keep track of storage capacity and
assigning storage to various applications accordingly. It saves time and user
efficiency. Regular monitoring of the physical health, failures, and outages can
help manage storage devices.
¢ Configuration, migration, provisioning, storage monitoring and reporting are the
management level tasks. Storage Resource Management (SRM) tools include
configuration tools, provisioning tools and measurement tools.
e Storage management tools include capacity management, monitoring, cloud
storage management and other devices and applications to manage SAN and NAS
systems. Storage management also includes backup and secondary storage tools
such as snapshots and copy data management.
1. Configuration tools : It handles the set-up of storage resources. These tools
help to organize and manage RAID devices.
2. Provisioning tools : It defines and control access to storage resources for
preventing a network user from being able to use any other user's storage.
3. Measurement tools : Analyse performance based on behavioural information
about a storage device.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 2-11 Data Storage and Cloud Computing

(2.2.2 | Storage Management Process

Storage management covers three area : Change management, performance and


capacity planning, tiering.
Change management : Process used to request, schedule, implement and evaluate
adjustments to the storage infrastructure is called change management.
Performance and capacity planning are used to measure the performance of a
system in terms of storage and utilization.
Storage management process is as follows :
a. Start the process : Data is stored according to its importance and access
frequency.
b. Data backup, restore and recovery operations : Data backups are made on a
periodic basis, also restore and recovery planes are made to guide restore and
recovery operations to bring backups online.
c. Storage resource management : Involves managing and maintaining those
resources on which the backed up data us stored like storage media.

ZX] Cloud Provisioning

Cloud provisioning is the allocation of a cloud provider's resources and services to


a customer. It is a key feature of the cloud computing model, relating to how a
customer procures cloud services and resources from a cloud provider. The
growing catalog of cloud services that customers can provision includes
infrastructure as a service, software as a service and platform as a service, in
public or private cloud environments.
Provisioning is the process by which a resource is prepared for use, reserved,
accessed, used and then released when the transaction is completed.
It is a key feature of the cloud computing model, relating to how a customer
procures cloud services and resources from a cloud provider.
The growing catalog of cloud services that customers can provision includes
infrastructure as a service, software as a service and platform as a service, in
public or private cloud environments.
Benefits of provisioning :
1. Continuous improvement of activity which relay on process measurement.
2. There is isolation of install, configure, build and customize task.

3. Ability to measure progress of all work


4 Assembly line approach to provisioning.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 2-12 Data Storage and Cloud Computing

Characteristics of provisioning :
1. Missing and incorrect information is resolved.
2. Product being deliver to customer before provisioning.
The characteristics of a provision are that it is a liability where there is uncertainty
as to either the timing of settlement or the amount to be settled.
When measuring a provision, the amount to be recognized should be the best
estimate of the consideration required to settle the present obligation at the end of
the reporting period.
The fact that it is difficult to measure a provision and that estimates have to be
used does not mean that the provision is not reliably measurable.
Types of cloud provisioning are as follows :
1. Dynamic provisioning : Cloud resources are deployed flexibly to match a
customer's fluctuating demands. The deployments typically scale up to
accommodate spikes in usage and scale down when demands decrease. The
customer is billed on a pay-per-use basis.
2. Self - provisioning : With user self-provisioning, also called cloud self-service,
the customer buys resources from the cloud provider through a web interface
or portal. This usually involves creating a user account and paying for
resources with a credit card.
3. Advance provisioning : Customer undertake contract with the provider for the
required services.

EX] File System


File system is an interface between secondary storage device like hard disk and
user application. Hard disk is a block oriented device.
The purpose of file systems is to maintain a consistent view of storage so that we
can effectively manage it. This is done in a way that allows the users to create
files and directories as well as delete, open, close, read, write and/or extend the
files on the device(s).
File systems also maintain security over the files by using access control lists for a
file.
File system is a structured data representation and a set of metadata that describe
the stored data. File system also support disk parition.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 2-13 Data Storage and Cloud Computing

EXXN FAT File System


File Allocation Table (FAT) file system is a simple file system originally designed
for small disks and simple folder structures.
Ususally the file system operates blocks, not sector. File system blocks are groups
of sectors that optimize storage addressing. Modern file systems generally use
block sizes from 1 up to 128 sectors (512 - 65536) bytes.
Windows Operating System use two major file systems : FAT and NTFS. Recently
released ReFS file system was developed by Microsoft as a new generation file
system for windows 8 servers.
The number in FAT12, FAT16 and FAT32 stands for the number of bits used in
file system block.
Linux uses Ext2, Ext3, Ext4 file system. Ext3 file system is just an extension to Ext2
that uses transactional file write operations with journal.

EXE] NTFS
NT File System (NTFS), which is also sometimes called the new technology file
system, is a process that the Windows NT operating system uses for storing,
organizing and finding files on a hard disk efficiently.
Formatting a volume with the NTFS file system results in the creation of several
system files and the Master File Table (MFT), which contains information about all
the files and folders on the NTFS volume.
NTES file system supports larger file sizes and hard drives and is more secure
than FAT. In NTFS, the size of cluster ranges from 512 bytes to 64 kilobytes.
NTES file system can help you to compress files and folders or the whole NTFS
partition to save disk space so you can make best use of it.

EXE] Cloud File System


Cloud storage is the abstraction, pooling and sharing of storage resources through
the internet. File storage is the dominant technology used on NAS systems and is
responsible for organizing data and representing it to users. Its hierarchical
structure allows us to navigate data from top to bottom easily, but increases
processing time.
A file system in the cloud is a hierarchical storage system that provides shared
access to file data. Users can create, delete, modify, read and write files and can
organize them logically in directory trees for intuitive access.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 2-14 Data Storage and Cloud Computing

Cloud file storage is most appropriate for unstructured data or semi - structured
data, such as documents, spreadsheets, presentations and other file - based data.
Cloud file storage is a storage service that is delivered over the Internet, billed on
a pay-per-use basis and has an architecture based on common file level protocols
such as Server Message Block (SMB), Common Internet File System (CIFS) and
Network File System (NFS).
Cloud File system should be scalable enough to adopt large organizations file
systems under different workloads with good performance requirements.
Cloud file storage is a method for storing data in the cloud that provides servers
and applications access to data through shared file systems. This compatibility
makes cloud file storage ideal for workloads that rely on shared file systems and
provides simple integration without code changes.
A file system in the cloud is a hierarchical storage system that provides shared
access to file data. Users can create, delete, modify, read, and write files and can
organize them logically in directory trees for intuitive access.
Cloud file storage can easily expand with multiple servers as per
Benefits of Cloud File Storage : demand thus it can used for large oragnizations also
Storing file data in the cloud delivers advantages in three key areas :
1. Scalability : Although not every cloud file storage solution can scale,
leveraging all the capabilities of the cloud, the most advanced solutions
provide the ability to start with the capacity and performance you need today
and grow your capacity as needed. No more over provisioning to try and
anticipate future needs.
2. Interoperability : Many existing applications require integration with shared
file services that follow existing file system semantics. Cloud file storage
solutions offer a distinct advantage as there is no new code to write to have
secure, shared file access.
local storage
3. Budget and Resources : Operating file services on-premises requires budget for
hardware, ongoing maintenance, power, cooling, and physical space. Cloud file
storage enables organizations to redeploy technical resources to other projects
that bring more value to the business.

1. Ghost File System :


Amazon Web Services (AWS) uses Ghost cloud File System (GFS). GFS run over
Amazon's EC2, simpleDB web service and S3.

GFS is elastic and cost efficient. It is highly secure. It can be mounted on server,
client or access files via web page.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 2-15 Data Storage and Cloud Computing

2. Hadoop File System :


¢ Handoop Distributed File System (HDFS) is a distributed file system inspired by
GFS that organizes files and stores their data on a distributed computing system.
eo Handoop is an open source software framework that supports data-instensive
distributed applications, licensed under the Apache v2 license. It provide software
framework for distributed processing of large datasets in real - time applications.
e Handoop Distributed File System is a block - structured file system where each
file is divided into blocks of a pre - determined size. These blocks are stored
across a cluster of one or several machines. block - level storage:- fixed size blocks madhe data
store kela jato
e Blocks are the nothing but the smallest continuous location on your hard drive
where data is stored. Similarly, HDFS stores each file as blocks which are scattered
throughout the Apache Hadoop cluster. The default size of each block is 128 MB
in Apache Hadoop 2.x (64 MB in Apache Hadoop 1.x) which you can configure as
per your requirement.

3. Kosmos File System (KFS) :


e KFS is an open source project written in C++ by search startup Kosmix.
e The Kosmos filesystem consists of three components :
a) One or multiple chunk servers that store the data on their own hard disks,
b) A metaserver that keeps an eye on the chunk servers, and
¢) An application that quickly gets rid of a single large file
e KFS first splits a file into handy 64 MB blocks. The filesystem distributes these
chunks evenly over all attached servers, aptly referred to as block or chunk
servers. The servers store the blocks on normal filesystems that belong to the host
operating systems.

1. Draw and explain Cloud File Systems GFS/HDFS. [RVR t a SURES We rt wy 3


2. Write a note on cloud file system with architectures. [|x VIER CY SIRS NCTC EL) 7 3

EX1 Cloud Data Stores SPPU : Dec.-16,17, March-19


eo Data store is a connection to a store of data, whether the data is stored in a
database or in one or more files. The data store may be used as the source of data
for a process.
e A datastore is a repository for storing, managing and distributing data sets on an
enterprise level.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 2-16 Data Storage and Cloud Computing

Datastore data may be structured, unstructured or in another electronic format.


Depending on the organization, a datastore may be classified as an application -
specific datastore, operational datastore or centralized datastore.

EXXH Distributed Data Store


A computer network where information is stored on more than one node, by
means of data replication it is termed as Distributed Data Store. It is used to refer
to a distributed database where users store information on a number of nodes or a
computer network in which users store information on a number of peer network
nodes.
Distributed data system is one in which files, scripts and images are stored in
more than one server or volumes rather than a single server as in traditional
systems.
Example : Google's Big Table, Amazon's Dynamo, windows Azure storage
These types of data store are non - relational databases that searches data quickly
over a large multiple node.

(2.4.2 | Data Store Types

EX¥X] BigTable
Bigtable is a distributed storage system that is used for managing and storing
structured data at Google.
Bigtable is designed to reliably scale to petabytes of data and thousands of
machine. Bigtable has multiple goals like applicability, high availability, scalability,
high performance.
It is used by approximate sixty Google project or product like Google Analytic,
Google Finance, Personalized search, Writely and Google Earth.
Bigtable is built on Google file system for storing the data for scheduling large
scale data processing. It stored data in form of rows, columns and timestamp that
means it maps with arbitrary string value like row key and colmn key as well as
timestamp.
Features :

1. Bigtable is a distributed storage system for managing strcutured data.


2. Bigtable uses the distributed Google File System (GFS) to store log and data
files.
3. A Bigtable cluster stores a number of tables. Each table consists of a set of
tablets and each tablet contains all data associated with a row range.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 2-17 Data Storage and Cloud Computing

4. Bigtable use a highly available and persistent distributed lock services called
Chubby.
5. Bigtable supports single - row transactions.
The map is indexed by a row key, column key and a timestamp ; each value in
the map is an un - interpreted array of bytes.
1. Row key :
The row keys in a table are arbitary strings. Every read or write of data under a
single row key is atomic.
Bigtable maintains data in lexicographic order by row key. The row range for a
table is dynamically parttioned.
Each row range is called a tablet, which is the unit of distribution and load
balancing.
2. Column key :
Column keys are grouped into sets called column families, which form the basic
unit of access control. All data stored in a column family is usually of the same
type.
A column family must be created before data can be stored under any column key
in that family; after a family has been created, any column key within the family
can be used.

XX] Dynamo
Dynamo is propriety key value structured storage system. It can act as database
and also distributed hash table.
Dynamo dynamically partitions a set of keys over a set of storage nodes
It is most powerful relational database available in WWW. Relational databases
have been used a lot in retail sites, to make visitors browse and search for product
easily.
Dynamo does not support replication.
Dynamo is used to manage the state of services that have very high reliability
requirements and need tight control over the tradeoffs between availability,
consistency, cost-effectiveness and performance.
There are many services on Amazon's platform that only need primary-key access
to a data store. For many services, such as those that provide best seller lists,
shopping carts, customer preferences, session management, sales rank, and product
catalog, the common pattern of using a relational database would lead to
inefficiencies and limit scale and availability. Dynamo provides a simple
primary-key only interface to meet the requirements of these applications.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 2-18 Data Storage and Cloud Computing

¢ Dynamo is a completely decentralized system with minimal need for manual


administration. Storage nodes can be added and removed from Dynamo without
requiring any manual partitioning or redistribution.
e Compared to Bigtable, Dynamo targets applications that require only key/value
access with primary focus on high availability where updates are not rejected even
in the wake of network partitions or server failures.
e Dynamo stores objects associated with a key through a simple interface; it exposes
two operations : get() and put().
¢ Dynamo treats both the key and the object supplied by the caller as an opaque
array of bytes. It applies a MD5 hash on the key to generate a 128-bit identifier,
which is used to determine the storage nodes that are responsible for serving the
key.
e Dynamo's partitioning scheme relies on consistent hashing to distribute the load
across multiple storage hosts. In consistent hashing , the output range of a hash
function is treated as a fixed circular space or "ring".
¢ Dynamo provides eventual consistency, which allows for updates to be propagated
to all replicas asynchronously.
¢ Dynamo uses vector clocks in order to capture causality between different versions
of the same object. A vector clock is effectively a list of (node, counter) pairs. One
vector clock is associated with every version of every object.
¢ In Dynamo, when a client wishes to update an object, it must specify which
version it is updating. This is done by passing the context it obtained from an
earlier read operation, which contains the vector clock information.
e In Dynamo, each storage node has three main software components: request
coordination, membership and failure detection, and a local persistence engine. All
these components are implemented in Java.

1. Write a short note on Dynamo. SPPU : Dec.-16, 17, Marks 5


6. Draw and explain architecture of Amazon Dynamo. JURE BERNE FE TR

[2.5 | Using Grids for Data Storage


e Grid computing is a distributed computing system where a group of computers
are connected to create and work as one large virtual computing power, storage,
database, application and service.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 2-19 Data Storage and Cloud Computing

Grid storage is a specific type of data storage that uses a series of interconnected
nodes to accommodate versatile and scalable storage systems.
Grid computing is based on physically distributed computer resources used
cooperatively to run one or more applications. The resources may be owned by
several organizations and the main goal is the efficient processing of specific
applications that need access to specific resources, for example, there may be data
resources held at various locations that must all be accessed by a particular
application.
Grid-Oriented Storage (GOS) is a dedicated data storage architecture which can
be connected directly to a computational grid to support advanced data bank
services and reservoirs for data that can be shared among multiple computers and
end users on the grid.
GOS is a successor of well - used Network-Attached Storage (NAS) products in
the Grid Computing.
GOS is designed to deal with long - distance, cross - domain and single - image
file operations, which is typical in Grid environments. GOS behaves like a file
server via the file - based GOS - FS protocol to any entity on the grid. Inspired by
the success of GridFTP, GOS - FS integrates a parallel stream engine and Grid
Security Infrastructure (GSI).

2.6 | Cloud Storage


Cloud storage solve problems of traditional storage. Cloud storage is a cloud
computing model that stores data on the internet through a cloud computing
provider who manages and operates data storage as a service.
For solving problem of traditional storage, following practices are used :
a) Huge and unpredictable growth
b) Cost and complexity of conventional storage
¢) Security.
Because cloud storage is handled on a pay - as - you - go basis, the platform is
built to handle rapid scaling as needed. In a local hardware scenario, additional
storage and hardware must be purchased, integrated and configured by the IT.
Cloud file storage providers have designed their infrastructure to handle client
needs of all shapes and sizes, meaning that they can adapt quickly in the case of a
client with an exploding user base.
Attributes of cloud storage :
1. Multitenancy : Multitenancy is when several different cloud customers are
accessing the same computing resources, such as when several different
companies are storing data on the same physical server.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 2-20 Data Storage and Cloud Computing

2. Resource pooling : Resource pooling allows cloud providers to pool large-scale


IT resources to serve multiple cloud consumers.
3. Scalable and elastic : Cloud storage services provide elasticity, which means
you can scale capacity as your data volumes increase or dial down capacity if
necessary.
On-demand self service
Pricing based usage : Cloud storage helps companies cut costs by eliminating
in-house storage infrastructure.

EXX] Data Management for Cloud Storage


Data management encompasses acquiring, storing, protecting and processing data
across an organization or business unit. It helps ensure that data is validated and
fully accessible to stakeholders when needed.
Data management in the cloud is part of a broader trend of data modernization.
In the past, cloud and data strategies were often managed separately. Today, they
should be part of a broader unified strategy that takes advantage of the cloud's
inherent elasticity, scalability and economies of scale to enable faster, more
powerful business insights and new capabilities.
The SNIA Cloud Storage Technologies Initiative (CSTI) is committed to the
adoption, growth and standardization of storage in cloud infrastructures. This
encompasses data services, orchestration and management, as well as the
promotion of portability of data in multi - cloud environments.
Fig. 2.6.1 shows cloud storage usage of SIRDM model.

Metadata

Read / Write data | Location System


User
Storage Data

Fr
HTTP Query, Access, Requirements
Application
Modify that drive
GET/PUT URIs specific
ACLs data services

ob
Fig. 2.6.1 Cloud storage usage of SIRDM model

SIRDM model uses three types of metadata : Storage system metadata, data
system metadata and user metadata.
Storage system metadata contain information about size, access time, owner and
ACL. Data system metadata contain information about encryption and retention
etc.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 2-21 Data Storage and Cloud Computing

2.6.2 | Cloud Data Management Interface

Cloud Data Management Interface is the interface for data management in cloud
computing. This functional interface can be used to retrieve, create, delete and
update data elements from any cloud.
It permits the client to find out the available capabilities in cloud storage. It
further helps in the management of the data and the containers place in them. It
also allows the accusation between the metadata and the containers.
CDMI features and benefits
a) CDMI adds powerful data management capabilities that include administration
and management of user accounts, security access, monitoring and billing
information.
b) CDMI clients can discover the capabilities of the cloud storage offering and use
this information to tailor the management of containers and the data placed in
them.
c¢) CDMI is an open international (ISO) standard designed to work alongside
OpenStack Swift and Amazon 53 Models.
The operations of the Cloud Data Management Interface are divided into two
types by international standards :
a) Those that use the CDMI content type in the HTTP body
b) Those who do not make the use of it
The CDMI can also be used by the management and administrative applications, it
defines the means to store and retrieve the data as well as manage the data. The
means by which data management is done is known as control path and the
means by which data retrieval is achieved is known as the data path.
CDMI can also manage the cloud storage properties and its capabilities are
discovered by the capabilities allowed to a client.

2.6.3 | Cloud Storage Requirement


1. Multi - tenancy
A multi - tenant cloud is a cloud computing architecture that allows customers to
share computing resources in a public or private cloud. Each tenant's data is
isolated and remains invisible to other tenants.
It allows multiple users to work in a software environment at the same time, each
with their own separate user interface, resources and services. The multitenant
application design was created to enable multiple users (tenants) to access the
same application logic simultaneously.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 2-22 Data Storage and Cloud Computing

2. Security :
Security policy establishes a set of security rules and regulations. Often, security
policies will further define how these rules and regulations are implemented and
enforced.
Securing data can be done using encryption, authentication and authorization.
3. Secure transmission channel
Following methods are used to secure network communications :
a. Virtual Private Network (VPN)
b. Hypertext Transfer Protocol Secure (HTTPS)
¢. Private Network
d. Transport Layer Security (TLS) and Secure Socket Layer (SSL)

4. Performance :

Cloud storage performance depends upon speed and latency.

EXX Provisioning Cloud Storage


Provisioning is the process of setting up IT infrastructure. It can also refer to the
steps required to manage access to data and resources and make them available to
users and systems.
Cloud provisioning means allocating a cloud service provider's resource to a
customer.
There are various cloud provisioning delivery models. Each model depends on the
types of resources or services an organization purchases, how and when the cloud
service provider delivers them and how customers pay for them. The three
models : Advanced, dynamic and user self - provisioning.
Cloud provisioning challenges :
1. Complex management and monitoring : Organizations may need several
provisioning tools to customize their cloud resources.
e Resource and service dependencies : Cloud applications and workloads often tap
into basic infrastructure resources, such as computing, networking and storage. But
public cloud service providers offer higher - level ancillary services like serverless
functions and Machine Learning (ML) and big data capabilities. Such services may
carry dependencies that can lead to unexpected overuse and surprise costs.
Policy enforcement : User cloud provisioning helps streamline requests and
manage resources but requires strict rules to make sure unnecessary resources are
not provided. That is time - consuming since different users require varying levels
of access and frequency.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 2-23 Data Storage and Cloud Computing

2.6.5 | Data Intensive Technologies for Cloud Computing

e Data - intensive systems encompass terabytes to petabytes of data. Such systems


require massive storage and intensive computational power in order to execute
complex queries and generate timely results.
it requires the more computational power to get the important insights from data
¢ Data Intensive Computing is defined as "a class of parallel computing applications
which use a data parallel approach to processing large volumes of data".
parallely data access karycha to process the data
1. Processing approach :
e Current data - intensive computing platforms use a "divide and conquer" parallel
processing approach combining multiple processors and disks in large computing
clusters connected using high - speed communications networks.
e It allows the data to be partitioned among the available computing resources and
processed independently to achieve performance and scalability based on the
amount of data.
e There are several important common characteristics of data intensive computing
systems that distinguish them from other forms of computing :
a) Data and applications or algorithms are co - located so that data movement is
minimized to achieve high performance in data intensive computing
b) Programming models that express the high - level operations on data such as
data flows are used and the runtime system transparently controls the
scheduling, execution, load balancing, communications and movement of
computation and data across the distributed computing cluster.
c) They provide reliability, availability and fault tolerance.
d) They are linearly scalable to handle large volumes of data.

2. System architecture :
MapReduce
¢ MapReduce is a parallel programming model proposed by Google. It aims at
supporting distributed computation on large datasets by using a large number of
computers with scalability and fault tolerance guarantees.
¢ During the map phase, the master node takes the input and divides it into sub -
problems, then distributes them to the worker nodes. Each worker node solves a
sub-problem and sends the intermediate results ready to be processed by reducer.
¢ During the reduce phase, intermediate results are processed by reduce function on
different worker nodes and the final results are generated.
e Map and reduce are two primitives in functional programming languages, such as
Lisp, Haskell, etc. A map function processes a fragment of a key - value pairs list
to generate a list of intermediate key - value pairs.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 2-24 Data Storage and Cloud Computing

¢ A reduce function merges all intermediate values associated with a same key and
produces a list of key - value pairs as output.
e Properties of Map Reduce :
a) Very large scale data.
b) MapReduce allows for distributed processing of the map and reduction
operations.
¢) MapReduce provides analytical capabilities for analyzing huge volumes of
complex data.
d) Number of maps tasks and reduce tasks are configurable.

High Performance Computing Cluster (HPCC) :


e HPCC is also known as DAS (Data Analytics Supercomputer), is an open source,
data - intensive computing system platform developed by LexisNexis Risk
Solutions. The DAS is a platform designed to refine, link and fuse large amounts
of data from disparate sources for complex analysis and queries.
o Architecturally, the DAS is a HPCC based on commodity server hardware, which
can be scaled up to thousands of processors to handle any amount of data and
runs on the Linux operating system.
e System software and middleware components were developed and layered on to
provide the execution environment, distributed file system and SOA interfaces
required to support data - intensive computing in an enterprise environment.

Cloud Storage from LANs to WANs


e The result of the migration from traditional distributed data storage to the cloud
computing based on data storage platform.

Cloud Characteristics

¢ Computer power is elastic, when it can perform parallel operations.


e Data is retained at an unknown host server.
e Data is duplicated often over distant location.

Distributed Data Storage


e A distributed data store is a computer network where information is stored on
more than one node, often in a replicated fashion. It is usually specifically used to
refer to either a distributed database where users store information on a number
of nodes or a computer network in which users store information on a number of
peer network nodes.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 2-25 Data Storage and Cloud Computing

1. Amazon dynamo :
Dynamo stores objects associated with a key through a simple interface; it exposes
two operations : get() and put().
The get(key) operation locates the object replicas associated with the key in the
storage system and returns a single object or a list of objects with conflicting
versions along with a context.
The put(key, context, object) operation determines where the replicas of the object
should be placed based on the associated key and writes the replicas to disk.
The context encodes system metadata about the object that is opaque to the caller
and includes information such as the version of the object. The context information
is stored along with the object so that the system can verify the validity of the
context object supplied in the put request.
Dynamo treats both the key and the object supplied by the caller as an opaque
array of bytes. It applies a MD5 hash on the key to generate a 128 - bit identifier,
which is used to determine the storage nodes that are responsible for serving the
key.
2. CouchDB :
Apache CouchDB is open source database software that has a document - oriented
NoSQL database architecture and is implemented in the concurrency - oriented
language Erlang; it uses JSON to store data.
CouchDB is ad-hoc and schema - free with a flat address space. CouchDB comes
with a developer - friendly query language and optionally MapReduce for simple,
efficient and comprehensive data retrieval.
Data is stored within JSON documents which can be accessed and its indices
queried via HTTP.
Indexing, transforming and combining of documents are performed through
JavaScript. Because it uses all of these web - friendly standards and technologies,
CouchDB works very well with web and mobile applications.
CouchDB is designed to store and report on large amounts of semi - structured,
document oriented data.
3. ThruDB
ThruDB is an open source database built on Apache's Thrift framework and is a
set of simple services such as scaling, indexing and storage which is used for
building and scaling websites.
It provides flexible, fast and easy-to-use services that simplify the management of
the modern web data layer and provides developers with features and tools most
web developers need. These features can be easily configured or turned off.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 2-26 Data Storage and Cloud Computing

eo Features
a) Multi - master replication
b) Built for horizontal scalability
¢) Incremental backups and redo logging
d) Multiple storage back - end client libraries for most languages
e) Simple and powerful search API
ThruDB provides web - scale data management by providing these services :
a) Thrucene - For Lucene - based indexing
b) Throxy - For partitioning and load balancing
¢) Thrudoc - For document storage
d) Thruqueue - For a persistent message queue service
e) Thrift - For cross - language services framework

[EX] Multiple Choice Questions


Q.1 NAS is a computer data storage server connected to a computer network
providing data access to a heterogeneous group of clients.

al file level b| directory level


c| root level d| all of these

Q.2 NAS support NFS and file system.

a| FAT b| NTFS
c| CIFS d| KFs

Q.3 Kosmos File System is an open source project written in by search startup
Kosmix.

al C b| Java
c| Machine language d C+

Q.4 ThruDB is an open-source database built on Apache's framework.

a| thrift b| network
c| file d| all of these

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 2-27 Data Storage and Cloud Computing

Q.5 Apache CouchDB is open-source database software that has a document-oriented


database architecture.

[a] sa [b] NosQL


PLSQL [d] Oracle

Q.6 Enterprise storage may include

[a] storage area network [b] network-attached storage


direct attached storage [d] all of these

Q.7 I/O requests to disk storage on a SAN are called

[al file 10s [b| SAN 1/Os


block I/Os [d| disk I/Os
Q.8 Which data storage technology offers the best performance ?

[a] saN [b] NAS


DAS [d| None

Q.9 NFS and CIFS are .

[a] file systems [b] operating systems


transport protocols [d] file access protocols

Q.10 What's the demerits of DAS ?

[a] Interconnect limited up to 10 km


bl Excessive network traffic

Distance limitations and slow speed


El Distance limitations and inability to share data with other servers

Answer Keys for Multiple Choice Questions :

Q1 a Q.2 c Q3 d Q4 a
Q.5 b Q.6 d Q.7 c Qs a

Q9 d | 0.10 d

aaa

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 2-28 Data Storage and Cloud Computing

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


UNIT - III

Virtualization in Cloud Computing

Syllabus
Introduction : Definition of Virtualization, Adopting Virtualization, Types of Virtualization,
Virtualization Architecture and Software, Virtual Clustering, Virtualization Application, Pitfalls of
Virtualization. Grid, Cloud and Virtualization : Virtualization in Grid, Virtualization in Cloud,
Virtualization and Cloud Security. Virtualization and Cloud Computing : Anatomy of Cloud
Infrastructure, Virtual infrastructures, CPU Virtualization, Network and Storage Virtualization.

Contents
3.1 Definition of Virtualization
3.2 Adopting Virtualization
3.3 Types of Virtualizations .................. March-19, -------------- Marks 5
3.4 Full Virtualization — .................. March-19, June-19- ---
-- - - Marks 5
3.6 Storage Virtualization
3.6 Virtual Clustering ~~ .................. March-20, -------------" Marks 5
3.7 Virtualization Application
3.8 Pitfalls of Virtualization
3.9 Grd, Cloud and Virtualization
3.10 Virtualization and Cloud Computing
3.11 Multiple Choice Questions

3-1)
Cloud Computing 3-2 Virtualization in Cloud Computing

3.1 Definition of Virtualization


Virtualization is a broad term that refers to the abstraction of resources across
many aspects of computing. For our purposes : One physical machine to support
multiple virtual machines that run in parallel. Virtualization is a frame work or
methodology of dividing the resources of computer into multiple execution
environments.
Virtualization is an abstraction layer that decouples the physical hardware from
the operating system to deliver greater IT resource utilization and flexibility. It
allows multiple virtual machines, with heterogeneous operating systems to run in
isolation, side-by-side on the same physical machine.
Virtualization means running multiple machines on a single hardware. The "Real"
hardware invisible to operating system. OS only sees an absiracted-out picture.
Only Virtual Machine Monitor (VMM) talks to hardware.
It is "a technique for hiding the physical characteristics of computing resources
from the way in which other systems, applications, or end users interact with
those resources. This includes making a single physical resource appear to function
as multiple logical resources; or it can include making multiple physical resources
appear as a single logical resource.”
Fig. 3.1.1 shows concept of virtualization.

Applications | Applications | Applications

Windows Unix Linux

Virtual Virtual Virtual


hardware hardware hardware

— Virtual —
Applications hardware Applications

Operating system Physical Instance, Single


Server which contains the
Hardware
CCPU , Memory , Secondary storage
all hardware resources

Fig. 3.1.1 Virtual machine

It is divided into two main categories :


1. Platform virtualization involves the simulation of virtual machines.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 3-3 Virtualization in Cloud Computing

2. Resource virtualization involves the simulation of combined, fragmented, or


simplified resources.
Following are the reasons for using virtualizations :
a) Virtual machines offer software developers isolated, constrained, test
environments.

b) The most important function of virtualization is the capability of running


multiple operating systems and applications on a single computer or server.
¢) Virtualization can usually improve overall application performance due to
technology that can balance resources, and provide only what the user needs.
d) It provides fault and error containment.
e) It helps in building secured computing platform.
f) Server virtualization P provides a way y to implement
p redundancy Cy without
purchasing additional hardware.

[EX] Adopting Virtualization


Virtualization is changing the IT environment where most companies have been
able to adopt it as a measure that allows for better utilization of hardware and
reduction of costs.
Virtualization is particularly valuable to small to medium businesses because it
lowers costs for hardware, and reduces systems administration and maintenance
costs because fewer servers are in operation.
Virtualization provides the server administrators a way to segment a large system
into smaller sub-systems. The server can be put to use more efficiently to meet the
different application needs for different users.
Virtualization speeds up resource delivery through centralized and automated
resource management. You can reduce the cost of a non-production environment
through virtualization tools.
Virtualization offers increased visibility and speed with which you can create a
security-focused, non-destructible environment to reduce certain risks.
Virtualizing the workstation environments can assist you with the simplification of
administration, regained control, and access to data.

Virtualization for enterprises deliver a seamless and standard quality user


experience and improvised control and security.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 3-4 Virtualization in Cloud Computing

[EX] Types of Virtualizations


e Virtualization is mainly used to emulate execution environment, storage and
network. Execution environment classified into two types : process level and
system level.
¢ Fig. 3.3.1 shows taxonomy of virtualization.

[How itis done? ( Technique ] [ Virtualization model|

[ | [ | [ |

— Emulation Application

++ Execution Fy Programming
1 environment Process level High-Level VM | ErEnEEE

© Storage L— Muliprogramming Operating system

Virtualization
“agg Network | Hardware-assisted
virtualization

1 Full virtualization

L— System level |— Hardware

— Paravirtualiztion

— Partial virtualization

Fig. 3.3.1 Taxonomy of virtualization

e Process level is implemented on top of an existing operating system.


e System level is implemented directly on hardware and do not or minimum
requirement of existing operating system.
Platform virtualization
¢ The creation of a virtual machine using a combination of hardware and software is
referred to as platform virtualization. Platform virtualization is performed on a
given hardware platform by "host" software, which creates a simulated computer
environment for its "guest" software.
¢ The "guest" software, which is often itself a complete operating system, runs just
as if it were installed on a stand-alone hardware platform. Typically, many such
virtual machines are simulated on a given physical machine.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 3-5 Virtualization in Cloud Computing

e For the "guest" system to function, the simulation must be robust enough to
support all the guest system's external interfaces, which may include hardware
drivers.
Resource virtualization
eo The basic concept of platform virtualization was later extended to the
virtualization of specific system resources, such as storage volumes, name spaces,
and network resources. Resource aggregation, spanning, or concatenation combines
individual components into larger resources or resource pools. For example : RAID
and volume managers combine many disks into one large logical disk.
e Virtual Private Network (VPN), Network Address Translation (NAT), and similar
networking technologies create a virtualized network namespace within or across
network subnets. Multiprocessor and multi-core computer systems often present
what appears as a single, fast processor.
e Application - level virtualization : It lets you emulate one application level
interface on another. Examples include JVM. Another example is WINE that lets
you run windows application on Linux or MAC by emulating the Win32 interface.
e Desktop virtualization : It supports various computing such as utility, testing,
security and development.

[EXXH Example : Wine

¢ Wine is a free and open-source compatibility layer that aims to allow application
software and computer games developed for Microsoft Windows to run on
Unix-like operating systems.
e Wine also provides a software library, named Winelib, against which developers
can compile Windows applications to help port them to Unix-like systems.
e Wine is developed with x86 architecture and does not emulate as a processor.
Fig. 3.3.2 shows wine, x86 based virtualization architecture.

Applications

Operating systems
ra) >

7 (9.4

Virtualization layer

I} U
[ x86 architecture |

Fig. 3.3.2 wine, x86 based virtualization architecture

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 3-6 Virtualization in Cloud Computing

x86 virtualization refers to hardware and software-based mechanisms to support


virtualization for processors based on the x86 architecture. Using a hypervisor, it
allows several operating systems to be run in parallel on an x86 processor and
resources to be distributed in an isolated and efficient manner between the
operating systems running in parallel.
1. FreeBSD : FreeBSD is a UNIX-like operating system. FreeBSD is used by
companies, Internet service providers, researchers, computer professionals,
students and home users all over the world in their work, education and
recreation.
2. Hive : Hive allows users to read, write, and manage petabytes of data using
SQL. Hive is built on top of Apache Hadoop, which is an open-source
framework used to efficiently store and process large datasets.
3. Nemesis : Nemesis is an operating system. Nemesis provides fine-grained
guaranteed levels of all system resources including CPU, memory, network
bandwidth and disk bandwidth. The OS has been built with the Multimedia in
mind, its sole purpose of existence was the delivery and performance of
multimedia content in the best way possible.

EER] Server Virtualization

Server virtualization is a software architecture that allows more than one server
operating system to run as a guest on a given physical server host. The concept of
Server Virtualization widely used in the IT infrastructure to minimizes the costs by
increasing the utilization of existing resources.
The ability to run multiple operating systems on a single physical system and
share the underlying hardware resources. Virtual machines provide a layer of
abstraction between the OS and the underlying hardware.
Creating multiple logical server OS instances on one physical piece of hardware.
All hardware drivers are virtualized and same virtual hardware regardless of
physical hardware.
Each virtual machine is completely independent of the others and doesn't realize’
it's virtualized.
Depending on the approach, server virtualization uses a number of different
components. These include :
1. A host machine, which is the physical server hardware where virtualization
occurs.
2. Virtual machines (VMs), which contain the assets that are abstracted from a
traditional server environment.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 3-7 Virtualization in Cloud Computing

3. A hypervisor, which is a specialized software that creates and maintains virtual


machines and can be run natively on bare metal servers or hosted on top of an
existing operating system.
Hypercalls, which are messages sent between para-virtualized hypervisors and
operating systems to share resources using an APL
Containers, which are unique user environments that are created in virtualized
operating systems. With a container engine, multiple containers can make use
of the same interfaces and shared libraries of the underlying host operating
system. Containers are often deployed inside of hypervisors or virtual
machines to offer an additional layer of isolation from the server's core host
operating system.
e Requirements of server virtualization :
1. Consolidation : It is common practice to dedicate each server to a single
application. If several applications only use a small amount of processing
power, the network administrator can combine several machines into one
server running multiple virtual environments.
Redundancy : Redundancy refers to running the same application on multiple
servers. It's a safety measure, if a server fails for any reason, another server
running the same application can take its place.
Legacy hardware : Server hardware will eventually become obsolete and
switching from one system to another can be difficult. In order to continue
offering the services provided by these outdated systems, sometimes called
legacy systems a network administrator could create a virtual version of the
hardware on modern servers.
Migration : Migration refers to moving a server environment from one place to
another. With the right hardware and software, it's possible to move a virtual
server from one physical machine in a network to another.
¢ Virtualization allows multiple operating system instances to run concurrently on a
single computer; it is a means of separating hardware from a single operating
system. Each "guest" OS is managed by a Virtual Machine Monitor (VMM), also
known as a hypervisor.
e Because the virtualization system sits between the guest and the hardware, it can
control the guests’ use of CPU, memory and storage, even allowing a guest OS to
migrate from one machine to another.
e By using specially designed software, an administrator can convert one physical
server into multiple virtual machines. Each virtual server acts like a unique
physical device, capable of running its own operating system.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 3-8 Virtualization in Cloud Computing

¢ Fig. 3.3.3 shows server virtualization after and before.

User application

Operating system

Computer hardware

(a) Before server virtualization

User User User User User


application application application application application

Operating system Operating system


Virtualization layer

Computer hardware

(b) After server virtualization

Fig. 3.3.3

e In server virtualization, the virtual servers are dedicated only to a particular task
for their betterment in performance. Every virtual server performs like a
distinctive physical device, that is capable of running its own operating system.
e Server virtualization is a cost-effective method that allows using resources
efficiently and provides web hosting services effectively utilizing existing resources
of IT infrastructure.
e By having each physical server divided into multiple virtual servers, server
virtualization allows each virtual server to act as a unique physical device. Each
virtual server can run its own applications and operating system. This process
increases the utilization of resources by making each virtual server act as a
physical server and increases the capacity of each physical machine.
e Types of server virtualization :
1. Full virtualization
2. Para-virtualization

3. OS level virtualization

o Benefits of server virtualization

1. Lower costs

2. Consolidation

3. Practice of redundancy

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 3-9 Virtualization in Cloud Computing

Disadvantages of server virtualization


1. Increase complexity of IT environment
2. Physical failures become "serious"
3. Bleed-over

EEE) Operating Level Virtualization

Operating-system-level virtualization is a server-virtualization method where the


kernel of an operating system allows for multiple isolated user-space instances,
instead of just one. Such instances, which are sometimes called containers and
software containers.
This refers to an abstraction layer between traditional OS and user applications.
This type of virtualization creates isolated containers on a single physical server
and the OS instances to utilize the hard-ware and software in data centers.
Containers behave like real servers. With containers you can create a portable,
consistent operating environment for development, testing, and deployment.
This virtualization creates virtual hosting environments to allocates hardware
resources among a large number of mutually distrusting users.
Operating-system-level virtualization usually imposes little to no overhead,
because programs in virtual partitions use the operating system’s normal system
call interface and do not need to be subjected to emulation or be run in an
intermediate virtual machine.
Operating system-level virtualization is not as flexible as other virtualization
approaches since it cannot host a guest operating system different from the host
one, or a different guest kernel.
Instead of trying to run an entire guest OS, container virtualization isolates the
guests, but doesn’t try to virtualize the hardware. Instead, you have containers for
each virtual environment.
With container-based technologies, you'll need a patched kernel and user tools to
run the virtual environments. The kernel provides process isolation and performs
resource management.
Why operating system level virtualization is required ?
Operating system level virtualization provides feasible solution for hardware level
virtualization issue. It inserts a virtualization layer inside an operating system
to partition a machine's physical resources.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 3-10 Virtualization in Cloud Computing

e It enables multiple isolated VMs within a single operating system kernel.


This kind of VM is often called a virtual execution environment (VE),
Virtual Private System (VPS), or simply container.
e From the user's point of view, virtual execution environment look like real servers.

e This means a virtual execution environment has its own set of processes, file
system, user accounts, network interfaces with IP addresses, routing tables,
firewall rules etc.
e Although VEs can be customized for different people, they share the
same operating system kernel. Therefore, OS-level virtualization is also called
single-OS image virtualization.
Challenges to cloud computing in OS level virtualization ?
e Cloud computing is transforming the computing landscape by shifting the
hardware and staffing costs of managing a computational center to third parties.
¢ Cloud computing has at least two challenges :
1. The ability to use a variable number of physical machines and virtual machine
instances depending on the needs of a problem. For example, a task may need
only a single CPU during some phases of execution but may need hundreds of
CPUs at other times.
2. It is related to slow operation of instantiating new virtual machine. Currently,
new virtual machines originate either as fresh boots or as replicates of a
template VM, unaware of the current application state. Therefore, to better
support cloud computing, a large amount of research and development should
be done.

Advantages of OS virtualization :
1. OS virtualization provide least overhead among all types of virtualization
solution.
They offer highest performance and highest density of virtual environment.
Low resource requirements.
4. High Scalability.

Disadvantage of OS virtualization :
1. They support only one operating system as base and guest OS in a single
server.
2. It supports library level virtualization.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 3-11 Virtualization in Cloud Computing

EE Para-Virtualization

Paravirtualization is a type of virtualization in which a guest operating system


(OS) is recompiled, installed inside a virtual machine (VM), and operated on top
of a hypervisor program running on the host OS.
Para-virtualization refers to communication between the guest OS and the
hypervisor to improve performance and efficiency.
Para-virtualization involves modifying the OS kernel to replace non-virtualizable
instructions with hyper-calls that communicate directly with the virtualization
layer hypervisor.
The hypervisor also provides hyper-call interfaces for other critical kernel
operations such as memory management, interrupt handling and time keeping.
Fig. 3.3.4 shows para-virtualization architecture.

Private Private Private


server server server

Operating system

Hardware

Fig. 3.3.4 Para-virtualization architecture

In Para-virtualization, the virtual machine does not necessarily simulate hardware,


but instead offers a special API that can only be used by modifying the "guest"
OS. This system call to the hypervisor is called a "hypercall” in Xen.

Xen is an open source para-virtualization solution that requires modifications to


the guest operating systems but achieves near native performance by collaborating
with the hypervisor.
Microsoft Virtual PC is a para-virtualization virtual machine approach. User-mode
Linux (UML) is another para-virtualization solution that is open source.

Each guest operating system executes as a process of the host operating system.
Cooperative Linux, is a virtualization solution that allows two operating systems
to cooperatively share the underlying hardware.
Linux-V server is an operating system-level virtualization solution for GNU/Linux
systems with secure isolation of independent guest servers.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 3-12 Virtualization in Cloud Computing

¢ The Linux KVM is virtualization technology that has been integrated into the
mainline Linux kernel . Runs as a single kernel loadable module, a Linux kernel
running on virtualization-capable hardware is able to act as a hypervisor and
support unmodified Linux and Windows guest operating systems.
e Para-virtualization shares the process with the guest operating system.
Problems with para-virtualization
1. Para-virtualized systems won't run on native hardware
2. There are many different para-virtualization systems that use different
commands, etc.
¢ The main difference between full virtualization and paravirtualization in Cloud is
that full virtualization allows multiple guest operating systems to execute on a
host operating system independently while paravirtualization allows multiple guest
operating systems to run on host operating systems while communicating .

1. Explain different levels of virtualization implementation with neat diagram. Also give example
of each. SPPU : March-19, In Sem, Marks 5

(3.4 Full Virtualization SPPU : March-19, June-19


e Full Virtualization doesn’t need to modify the host OS; it relies upon binary
translation to trap and to virtualize certain sensitive instructions.
¢ Fig. 3.4.1 shows full virtualization.
eo VMware Workstation applies full p
virtualization, which uses binary ( \
translation to automatically modify — —
x86 software on-the-fly to replace ( Application ) ( Application )
critical instructions. - -
Guest Operating Guest Operating
e Normal instructions can run ( system | ( system |
directly on the host OS. This is
done to increase the performance
overhead - normal instructions are ( Virtualization layer )
carried out in the normal manner, hy Hardware ’
but the difficult and precise
executions are first discovered £2 > OY > 4
using a trap and executed in a \ J
virtual manner. Fig. 3.4.1 Full virtualization

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 3-13 Virtualization in Cloud Computing

This is done to improve the security of the system and also to increase the
performance.
Host based virtualization :
Virtualization implemented in a host computer rather than in a storage subsystem
or storage appliance.
Virtualization can be implemented either in host computers, in storage subsystems
or storage appliances, or in specific virtualization appliances in the storage
interconnect fabric.
The guest OS are installed and run on top of the virtualization layer. Dedicated
applications may run on the VMs. Certainly, some other applications can also run
with the host OS direcily.
Advantages of host-based architecture :
1. The user can install this VM architecture without modifying the host OS.
2. The host-based approach appeals to many host machine configurations.

3.4.1] Memory Virtualization

Memory virtualization features allow abstraction isolation and monitoring of


memory on a per Virtual Machine (VM) basis. These features may also make live
migration of VMs possible, add to fault tolerance, and enhance security.
Example features include Direct Memory Access (DMA) remapping and Extended
Page Tables (EPT), including their extensions: accessed and dirty bits, and fast
switching of EPT contexts.
The VMkernel manages all machine memory. The VMkernel dedicates part of this
managed machine memory for its own use. The rest is available for use by virtual
machines.
Virtual machines use machine memory for two purposes : each virtual machine
requires its own memory and the VMM requires some memory and a dynamic
overhead memory for its code and data.
The virtual memory space is divided into blocks, typically 4KB, called pages. The
physical memory is also divided into blocks, also typically 4KB.
When physical memory is full, the data for virtual pages that are not present in
physical memory are stored on disk. ESX/ESXi also provides support for large
pages.
The VMM is responsible for mapping the guest physical memory to the actual
machine memory.
Each page table of a guest OS has a page table allocated for it in the VMM. The
page table in the VMM which handles all these is called a shadow page table.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 3-14 Virtualization in Cloud Computing

e As it can be seen all this process is nested and inter-connected at different levels
through the concerned address.
e If any change occurs in the virtual memory page table or TLB, the shadow page
table in the VMM is updated accordingly.

3.4.2] I/O Virtualization

e I/O Virtualization involves managing of the routing of I/O requests between


virtual devices and shared physical hardware.
e The there are three ways to implement this are full device emulation, para-VZ and
direct 1/0.
e I/O virtualization features facilitate offloading of multi-core packet processing to
network adapters as well as direct assignment of virtual machines to virtual
functions, including disk I/O.

e Examples include Virtual Machine Device Queues (VMDQ), Single Root I/O
Virtualization.
e Fig. 3.4.2 shows I/O virtualization.
VM Host
.
Applications ~~ Virtual Machines

[App] ~~
4
/
[Aer] | App |
Guest OS
) Hypervisor
Virtual-to-physical
translation
1/0 stack

1/0 stack

Interpose/transform
=a

oO

e.g. Log, encrypt


[0]
[+]
3
[3
jy

1/0 scheduler

Virtual hardware N\ Physical hardware

Emulated Local
disk device disk device NTC
\ > \.

Fig. 3.4.2 I/O virtualization

1. Full device emulation : This process emulates well-known and real-world


devices. All the functions of a device or bus infrastructure such as device
enumeration, identification, interrupts etc. are replicated in the software, which
itself is located in the VMM and acts as a virtual device. The I/O requests are
trapped in the VMM accordingly.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 3-15 Virtualization in Cloud Computing

2. Para-virtualization : This method of I/O VZ is taken up since software


emulation runs slower than the hardware it emulates. In para-VZ, the frontend
driver runs in Domain-U; it manages the requests of the guest OS. The
backend driver runs in Domain-0 and is responsible for managing the real 1/0
devices. This methodology (para) gives more performance but has a higher
CPU overhead.

3. Direct I/O virtualization : This lets the VM access devices directly; achieves
high performance with lower costs. Currently, it is used only for the
mainframes.

3.4.3 | Difference between Full and Para Virtualization

Sr. No. Full Virtualization Para Virtualization

1 Full Virtualization relies upon binary Para-Virtualization refers to communication


translation to trap and to virtualize between the guest OS and the hypervisor to
certain sensitive instructions. improve performance and efficiency.
Example : VMware Example : Xen architecture
2. Full Virtualization doesn't need to Para-Virtualization involves modification of
modify the host OS. OS kernel.

3. Normal instructions can run directly on Para-virtualized systems won't run on


the host OS. native hardware.

4. Full Virtualization uses binary Para-Virtualization uses hyper - calls.


translation and direct execution.

5. Performance is good. Performance is better in certain cases.

6. Guest software does not require any Hardware is not simulated and the guest
modification since the undelying software run their own isolated domains.
hardware is fully simulated.

1. Explain full and para virtualization with examples. SPPU : March-19, In Sem, Marks 5

2. Explain the following : i) CPU virtualization ii) Memory virtualization.


SPPU : June-19, End Sem, Marks 5

[EX storage Virtualization


e Storage virtualization today refers to a wide variety of products and technologies,
from the simplest file systems all the way up-to the cutting edge storage
abstraction layers that are capable of managing terabytes of heterogeneous storage
spread across the world under a single coherent management framework.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 3-16 Virtualization in Cloud Computing

e Storage virtualization refers to the abstraction of storage systems from applications


or computers. It is a foundation for the implementation of other technologies, such
as thin provisioning and data protection, which are transparent to the server.
e Storage virtualization provides the ability to pool storage systems into a
consolidated, shared capacity that can be managed from a central point of control.
e Example of storage virtualizations are host-based volume management, LUN
creation, tape storage virtualization and disk addressing.
e Storage virtualization has the following characteristics :
1. The availability of logical volumes separate from physical hard disk constraints
2. The capability of abstracting multivendor storage devices into one group and
reallocating storage space independently of size or physical location
3. The capability of having automated storage optimization and management.
e Fig. 3.5.1 shows virtualized storage environment.

Server Server Server Server Server

Virtual
volumes

Virtualization layer
LOA)

Heterogeneous physical storage

Fig. 3.5.1 Storage virtualization

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 3-17 Virtualization in Cloud Computing

e Top level servers assigned one virtual volume, which is currently in use by an
application. These virtual volumes are mapped to the actual storage in the arrays.
When an 1/0 is sent to a virtual volume, it is redirected through the virtualization
at the storage network layer to the mapped physical array
e Primary types of storage virtualizations are block level virtualization and file
virtualization.
e Block level virtualization : It separates physical and logical storage. File
virtualization optimes use of server and storage consolidation.
¢ Block-based : Block-based storage virtualization is the most common type of
storage virtualization being practiced across organizations. It identifies all available
blocks on individual media/path irrespective of location or vendor, and then the
engine leaves that data in the physical position and maps the address to a virtual
storage device.
¢ File-based : File-level virtualization works over NAS devices. It has a challenge of
its own because managing different NAS devices can be tedious work. Managing
multiple appliances is time-consuming and costly. NAS devices require individual
management, and users need to know the physical pathname to access a file.
Migration of data from old to new NAS devices also remains a challenge as it
results in downtime, leading to additional cost to the company.
e Currently there are three methods of storage virtualization :
1. Server-based virtualization : This method places a management program on the
host system and has the benefit of leveraging the SAN asset as it is.
2. Fabric-based virtualization : This can be done via network switches or
appliance servers. In both instances, independent appliances, such as switches,
routers, and dedicated servers, are placed between servers and storage and
have a storage virtualization function. The purpose behind this is to reduce the
impact on the existing SAN and servers.
3. Storage array-based virtualization : This is a virtualization implemented at the
storage-system level.
o Benefits to storage virtualization :
1. Data is stored in more convenient locations away from the specific host.
2. The storage devices are able to perform advanced functions like de-duplication,
replication, thin provisioning and disaster recovery functionality.
3. By abstracting the storage level, IT operations can become more flexible in how
storage is partitioned, provided and protected.
4. Improved physical resource utilization.
5. Lower total cost of ownership : Virtualized storage allows more to be done
with the same or less storage.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 3-18 Virtualization in Cloud Computing

[EXXE Network Virtualization


Network virtualization refers to the technology that enables partitioning or
aggregating a collection of network resources and presenting them to various users
in a way that each user experiences an isolated and unique view of the physical
network.
Network virtualization creates virtual networks whereby each application sees its
own logical network independent of the physical network.
A virtual LAN (VLAN) is an example of network virtualization that provides an
easy, flexible, and less expensive way to manage networks.
VLANs make large networks more manageable by enabling a centralized
configuration of devices located in physically diverse locations.
Fig. 3.5.2 shows network virtualization.
. . A Workload Workload Workload
Consider a company in which the
users of a department are separated
over a metropolitan area with their
L2,L3,L4-7 Network services )
resources centrally located at one (
office.
Virtual Virtual Virtual
In a typical network, each location network 1 network 2 network 3
has its own network connected to the E—
others through routers. When network ( Network virtualization platform )
packets cross routers, latency Requirement : IP transport
influences network performance.
With VLANs, users with similar
access requirements can be grouped
together into the same virtual
network. This setup eliminates the
need for network routing. RS
Physical network
As a result, although users are
physically located at disparate Fig. 3.5.2 Network virtualization
locations, they appear to be at the same location accessing resources locally.
In addition to improving network performance, VLANs also provide enhanced
security by isolating sensitive data from the other networks and by restricting
access to the resources located within the networks.
Network virtualization decouples the roles of the traditional Internet service
providers (ISPs) into infrastructure providers (InPs) and service providers (SPs).

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 3-19 Virtualization in Cloud Computing

e Benefits :
1. Reduces the number of physical devices needed.
2 Easily segment networks.
3. Permits rapid change / scalability and agile deployment.
4 Security from destruction of physical devices.

[EX] virtual Clustering also from notebook


e A computer cluster is a set of connected computers (nodes) that work together as
if they are a single machine. All processor machines share resources such as a
common home directory and have a software such as a Message Passing Interface
(MPI) implementation installed to allow programs to be run across all nodes
simultaneously.
e Computer clusters are often used for cost-effective High Performance Computing
(HPC) and High Availability (HA) by businesses of all sizes. A computer cluster
help to solve complex operations more efficiently with much faster processing
speed, better data integrity than a single computer and they only used for
mission-critical applications.
Characteristics Virtual Cluster : Virtual Machines are used to form the virtual Cluster.

1. Virtual machine or physical machine is used as virtual cluster nodes. Multiple


VM running with different types of OS can be deployed on the same physical
node. server
Virtual machine runs with guest operating system. Host OS and VM OS are
different but it manages the resources in the physical machine.
Virtual machine can be replicated in multiple servers and it support distributed
parallelism, fault tolerance and disaster recovery.
4. Number of nodes of a virtual cluster may change accordingly.
5. It virtual machine failes, it can not affect the host machine.
e Virtual cluster is managed by four ways :
1. We can use a guest-based manager, by which the cluster manager resides
inside a guest OS. Ex. : A Linux cluster can run different guest operating
systems on top of the Xen hypervisor.
We can bring out a host-based manager which itself is a cluster manager on
the host systems. Ex. : VMware HA (High Availability) system that can restart
a guest system after failure.
An independent cluster manager, which can be used on both the host and the
guest - making the infrastructure complex.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 3-20 Virtualization in Cloud Computing

4. Finally, we might also use an integrated cluster (manager), on the guest and
host operating systems; here the manager must clearly distinguish between
physical and virtual resources.

1. Explain in brief virtual clusters and resource management.


SPPU : March-20, In Sem, Marks 5

3.7 Virtualization Application

Virtualization at the application level virtualizes an application as a VM. On a


traditional OS, an application often runs as a process. Therefore, application-level
virtualization is also known as process-level virtualization.
A fully virtualized application is not installed in the traditional sense, although it
is still executed as if it were. The application behaves at runtime like it is directly
interfacing with the original operating system and all the resources managed by it,
but can be isolated to varying degrees.
Full application virtualization requires a virtualization layer. Application
virtualization layers replace part of the runtime environment normally provided
by the operating system.
The layer intercepts all disk operations of virtualized applications and
transparently redirects them to a virtualized location, often a single file.
The application remains unaware that it accesses a virtual resource instead of a
physical one. Since the application is now working with one file instead of many
files spread throughout the system, it becomes easy to run the application on a
different computer and previously incompatible applications can be run
side-by-side.
The most popular approach is to deploy High Level Language (HLL) VMs. Here
the virtualization layer sits as an application program on top of the operating
system, and the layer exports an abstraction of a VM that can run programs
written and compiled to a particular abstract machine definition. Any program
written in the HLL and compiled for this VM will be able to run on it.
Benefits :
1. Application virtualization uses fewer resources than a separate virtual machine.
2. Application virtualization also enables simplified operating system migrations.
3. Applications can be transferred to removable media or between computers
without the need of installing them, becoming portable software.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 3-21 Virtualization in Cloud Computing

¢ Limitations :
1. Not all computer programs can be virtualized
2. Lower performance

[EX] Pitfalls of Virtualization


a) Pros
1. Data center and energy-efficiency savings : As companies reduce the size of
their hardware and server footprint, they lower their energy consumption.
2. Operational expenditure savings : Once servers are virtualized, your IT staff
can greatly reduce the ongoing administration and management of manual
work.
3. Reduced costs : It reduced cost of IT infrastructure.
4. Data does not leak across virtual machine.
5. Virtual machine is completed isolated from host machine and other virtual
machine.
6. Simplifies resource management by pooling and sharing resources.
7. Significantly reduce downtime.
8. Improved performance of IT resources.

1. Not all hardware or software can be virtualized.

2. Not all servers are applications are specifically designed to be


virtualization-friendly.

[EX] Grid, Cloud and Virtualization


EXXN Virtualization in Grid
eo A computational grid is a hardware and software infrastructure that provides
dependable, consistent, pervasive and inexpensive access to high-end
computational facilities.
e Grid computing is a distributed computing system where a group of computers
are connected to create and work as one large virtual computing power, storage,
database, application and service.
e Grid computing can be used in a variety of ways to address various kinds of
application requirements. The three primary types of grids are given below.
1. Computational grid : A computational grid is focused on setting aside
resources specifically for computing power. In this type of grid, most of the
machines are high-performance servers.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 3-22 Virtualization in Cloud Computing

2. Scavenging grid : A scavenging grid is most commonly used with large


numbers of desktop machines. Machines are scavenged for available CPU
cycles and other resources. Owners of the desktop machines are usually given
control over when their resources are available to participate in the grid.
3. Data grid : A data grid is responsible for housing and providing access to data
across multiple organizations. Users are not concerned with where this data is
located as long as they have access to the data. For example, you may have
two universities doing life science research, each with unique data. A data grid
would allow them to share their data, manage the data and manage security
issues such as who has access to what data.

3.9.2 | Virtualization in Cloud


1. The cloud computing adoption model
e Cloud Adoption is a strategic move by organizations of reducing cost, mitigating
risk and achieving scalability of data base capabilities. Fig. 3.9.1 shows cloud
computing adoption model.

Level 5: Achieve dynamic sharing of application


workload, capacity arbitrage and
self-service application provisioning

Level 4 : Select cloud environmnet and


begin broad-based deployments; Cloud
manual provisioning and load exploitation
balancing
Level 3 : Lay foundation for
Cloud scalable application
foundations architecture
Level 2 : Experiment in
amazon EC2; define
reference architecture Cloud
experimentation

Level 1: Virtualize
infrastructure and
Virtualization applications

Fig. 3.9.1 Cloud computing adoption model

e Cloud adoption model consists of following layers :


a) Hyper cloud : It provides dynamic sharing and self-service application.
b) Cloud foundations : It performs load balancing, deployments etc.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 3-23 Virtualization in Cloud Computing

¢) Cloud exploitation : Where foundations for scalable application architecture are


carried out.
d) Cloud experimentation : Various architectures are experimented.
e) Virtualization : Infrastructure and applications are virtualized.

3.9.3 | Difference between Cloud and Virtualization

Sr. No. Virtualization Cloud Computing

1. Virtualization is the process of creating a Cloud computing means storing and


virtual environment on an existing server accessing data and programs over the
to run your desired program, without Internet instead of your computer's hard
interfering with any of the other services drive.
provided by the server or host platform to
other users.
Location of virtual machine is on a Location of virtual machine is on any host.
specific host.
Instance storage is persistent. Instance storage is shortly lived.
Virtualization uses customizable VM Cloud computing uses standard VM
resource like CPU and RAM. resource like CPU and RAM
Recovery from failures: attempt to recover Recovery from failures : Discard instance
failed VM. spin up new one.

3.9.4 | Virtualization and Cloud Security

Cloud computing security challenges fall into three broad categories :


1. Data protection : Securing your data both at rest and in transit.
2. User authentication : Limiting access to data and monitoring who accesses the
data.
3. Disaster and data breach : Contingency planning.
Data protection : Data needs to be encrypted at all times, with clearly defined
roles when it comes to who will be managing the encryption keys.
User authentication : Data resting in the cloud needs to be accessible only by those
authorized to do so, making it critical to both restrict and monitor who will be
accessing the company's data through the cloud. In order to ensure the integrity of
user authentication, companies need to be able to view data access logs and audit
trails to verify that only authorized users are accessing the data.
Contingency planning : With the cloud serving as a single centralized repository
for a company's mission-critical data, the risks of having that data compromised
due to a data breach or temporarily made unavailable due to a natural disaster are
real concerns.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 3-24 Virtualization in Cloud Computing

e If information is encrypted while passing through the cloud, who controls the
encryption/decryption keys ? Is it the customer or the cloud vendor ? Most
customers probably want their data encrypted both ways across the Internet using
Secure Sockets Layer protocol.
e They also most likely want their data encrypted while it is at rest in the cloud
vendor's storage pool. Be sure that you, the customer, control the
encryption/decryption keys, just as if the data were still resident on your own
servers.
e Data integrity means ensuring that data is identically maintained during any
operation.
¢ Cloud-based services will result in many mobile IT users accessing business data
and services without traversing the corporate network. This will increase the need
for enterprises to place security controls between mobile users and cloud-based
services.
¢ Placing large amounts of sensitive data in a globally accessible cloud leaves
organizations open to large distributed threats, attackers no longer have to come
onto the premises to steal data, and they can find it all in the one "virtual"
location.
e Virtualization efficiencies in the cloud require virtual machines from multiple
organizations to be co-located on the same physical resources. Although traditional
data center security still applies in the cloud environment, physical segregation
and hardware-based security cannot protect against attacks between virtual
machines on the same server.
¢ Operating system and application files are on a shared physical infrastructure in a
virtualized cloud environment and require system, file, and activity monitoring to
provide confidence and auditable proof to enterprise customers that their resources
have not been compromised or tampered with.
e In the cloud computing environment, the enterprise subscribes to cloud computing
resources, and the responsibility for patching is the subscriber's rather than the
cloud computing vendor's.
e The need for patch maintenance vigilance is imperative. Lack of due diligence in
this regard could rapidly make the task unmanageable or impossible, leaving you
with "virtual patching” as the only alternative.
e Confidentiality : Confidentiality refers to limiting information access. Sensitive
information should be kept secret from individuals who are not authorized to see
the information.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 3-25 Virtualization in Cloud Computing

¢ In cloud environments, confidentiality primarily pertains to restricting access to


data in transit and storage.
e Integrity can extend to how data is stored, processed, and retrieved by cloud
services and cloud-based IT resources.
¢ Some common cloud security threats include :
a) Risks of cloud-based infrastructure including incompatible legacy IT
frameworks, and third-party data storage service disruptions.
b) Internal threats due to human error such as misconfiguration of user access
controls.
c) External threats caused almost exclusively by malicious actors, such as
malware, phishing, and DDoS attacks.

[EXT] virtualization and Cloud Computing


[EETXE Anatomy of Cloud Infrastructure
¢ Cloud anatomy can be simply defined as the structure of the cloud. Anatomy can
be considered as a part of architecture. Fig. 3.10.1 shows cloud anatomy.

Application

Platform

Infrastructure

Virtualization

Server/storage/datacenters

Fig. 3.10.1 Cloud anatomy

1. Application : The upper layer is the application layer. In this layer, any
applications are executed.
2. Platform : This component consists of platforms that are responsible for the
execution of the application. This platform is between the infrastructure and
the application.
3. Infrastructure : The infrastructure consists of resources over which the other
components work. This provides computational capability to the user.
4, Virtualization : Virtualization is the process of making logical components of
resources over the existing physical resources. The logical components are
isolated and independent, which form the infrastructure.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 3-26 Virtualization in Cloud Computing

5. Physical hardware : The physical hardware is provided by server and storage


units.

[EXT®] Virtual Infrastructures


A virtual infrastructure allows you to utilise the IT capabilities of physical
resources as software that can be used across multiple platforms. These resources
are shared across multiple virtual machines (VMs) and applications for maximum
efficiency, creating a virtual infrastructure.
Virtual Infrastructure consists of the following components :
a) Bare-metal hypervisors to enable full virtualisation of each x86 computer.
b) Virtual infrastructure services such as resource management and consolidated
backup to optimise available resources among virtual machines.
c) Automation solutions that provide special capabilities to optimise a particular
IT process such as provisioning or disaster recovery.
Cloud computing provides virtual infrastructures which provide facility for data
storage and computing power without direct management by users.

[EXTEY CPU virtualization


Certain processors such as Intel VI provide hardware assistance for CPU
virtualization.
When using this assistance, the guest can use a separate mode of execution called
guest mode. The guest code, whether application code or privileged code, runs in
the guest mode.
On certain events, the processor exits out of guest mode and enters root mode.
The hypervisor executes in the root mode, determines the reason for the exit, takes
any required actions, and restarts the guest in guest mode.
When you use hardware assistance for virtualization, there is no need to translate
the code. As a result, system calls or trap-intensive workloads run very close to
native speed.
Some workloads, such as those involving updates to page tables, lead to a large
number of exits from guest mode to root mode. Depending on the number of such
exits and total time spent in exits, this can slow down execution significantly.
CPU virtualization features enable faithful abstraction of the full prowess of Intel
CPU to a virtual machine.
All software in the VM can run without any performance, as if it was running
natively on a dedicated CPU. Live migration from one Intel CPU generation to
another, as well as nested virtualization, is possible.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 3-27 Virtualization in Cloud Computing

Er Network and Storage Virtualization

Network virtualization refers to the technology that enables partitioning or


aggregating a collection of network resources and presenting them to various users
in a way that each user experiences an isolated and unique view of the physical
network.
Network virtualization creates virtual networks whereby each application sees its
own logical network independent of the physical network.
A virtual LAN (VLAN) is an example of network virtualization that provides an
easy, flexible, and less expensive way to manage networks.
VLANs make large networks more manageable by enabling a centralized
configuration of devices located in physically diverse locations.
Fig. 3.10.2 shows network virtualization.

| Workload | | Workload | | Workload

L2,L3,L4-7 Network services

Virtual Virtual Virtual


network 1 network 2 network 3

( Network virtualization platform )

Requirement : IP transport

Physical network

Fig. 3.10.2 Network virtualization

Consider a company in which the users of a department are separated over a


metropolitan area with their resources centrally located at one office.
In a typical network, each location has its own network connected to the others
through routers. When network packets cross routers, latency influences network
performance.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 3-28 Virtualization in Cloud Computing

With VLANS, users with similar access requirements can be grouped together into
the same virtual network. This setup eliminates the need for network routing.

As a result, although users are physically located at disparate locations, they


appear to be at the same location accessing resources locally.

In addition to improving network performance, VLANs also provide enhanced


security by isolating sensitive data from the other networks and by restricting
access to the resources located within the networks.
Network virtualization decouples the roles of the traditional Internet Service
Providers (ISPs) into Infrastructure Providers (InPs) and Service Providers (SPs)

Benefits :
1. Reduces the number of physical devices needed.

2. Easily segment networks.


3. Permits rapid change/scalability and agile deployment.

4. Security from destruction of physical devices.

[EXE] Multiple Choice Questions


Ql Which of the following type of virtualization is also characteristic of cloud
computing ?

a| Storage b| Application
c| CPU d| All of the mentioned

Q.2 Which of the following network resources can be load balanced ?

a| Connection through intelligent switches |b| DNS


c| Storage resources d| All of these

Q3 Each guest OS is managed by a virtual machine monitor also known as .

a| server b| hypervisor
c| storage d| none

Q4 is the process of making logical components of resources over the existing


physical resources.

a| Virtualization b| Cloud computing


c| Storage d| Loading

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 3-29 Virtualization in Cloud Computing

Q.5 Which of the following are types of server virtualization ?

[al Full virtualization [b] Para - virtualization

OS level virtualization [d] All of these

Q.6 Which of the following type of virtualization is also characteristic of cloud


computing ?

[a] Storage [b] Application


CPU [d] All of these

Q.7 What is the solution for full virtualization ?

[a] Processor [b] Application


Desktop [d| Hardware

Q.8 The creation of a virtual machine using a combination of hardware and software is
referred to as virtualization.

[a] system [b] CPU


machine [d| platform

Q.9 Library-level virtualization is also known as user-level interface.

[a] software [b] user


application [d] application binary

Answer Keys for Multiple Choice Questions :

0.1 c Q.2 d 0.3 b Q4 a

Q5 d Q.6 d Q7 d Qs d

Q9 d

aaa

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 3-30 Virtualization in Cloud Computing

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


UNIT - IV

Cloud Platforms and Cloud Applications

Syllabus
Amazon Web Services (AWS) : Amazon Web Services and Components, Amazon Simple DB,
Elastic Cloud Computing (EC2), Amazon Storage System, Amazon Database services (Dynamo
DB). Microsoft Cloud Services : Azure core concepts, SQL Azure, Windows Azure Platform
Appliance. Cloud Computing Applications : Healthcare : ECG Analysis in the Cloud, Biology :
Protein Structure Prediction, Geosciences : Satellite Image Processing, Business and Consumer
Applications : CRM and ERP, Social Networking, Google Cloud Application : Google App Engine.
Overview of OpenStack architecture.

Contents
4.1 Amazon Web Services .................. Dec.-19, --------------- Marks 8
4.2 Elastic Cloud Computing ................. June-19, Dec.-19, -------- Marks 9
4.3 Amazon Storage System
4.4 Amazon Database Services
4.5 Microsoft Cloud Services : Azure
4.6 Cloud Computing Applications
4.7 Google Cloud Application : Google App Engine
4.8 Overview of OpenStack Architecture
4.9 Multiple Choice Questions

(4-1)
Cloud Computing 4-2 Cloud Platforms and Cloud Applications

EXE Amazon Web Services


Amazon Web Services (AWS) is a cloud computing platform from Amazon that
provides customers with a wide array of cloud services.
Amazon first debuted its Amazon Web Services in 2006 as a way to enable the use
of online services by client-side applications or other web sites via HTTP, REST or
SOAP protocols.
Amazon bills customers for Amazon AWS based on their usage of the various
Amazon Web Services.
In 2012, Amazon launched the AWS Marketplace to accommodate and grow the
emerging ecosystem of AWS offerings from third-party providers that have built
their own solutions on top of the Amazon Web Services platform.
The AWS Marketplace is an online store for Amazon Web Services customers to
find, compare and begin using AWS software and technical services.
Amazon Web Services is a secure cloud services platform, offering compute
power, database storage, content delivery and other functionality to help
businesses scale and grow.
In 2017, AWS comprised more than 90 services spanning a wide range including
computing, storage, networking, database, analytics, application services,
deployment, management, mobile, developer tools, and tools for the Internet of
Things.
Today, Amazon Web Services provides a highly reliable, scalable, low-cost
infrastructure platform in the cloud that powers hundreds of thousands of
businesses in 190 countries around the world.
In 2016 AWS partnered with Digital Currency Group to create a laboratory
environment allowing companies to experiment with block chain technologies.
In January 2018, Amazon launched an autoscaling service on AWS.
What is Amazon Web Services ?
Amazon Web Services (AWS) is a collection of remote computing services (web
services) that together make up a cloud computing platform, offered over the
Internet by Amazon.com.
The AWS Cloud infrastructure is built around Regions and Availability Zones
(AZs). A Region is a physical location in the world where we have multiple AZs.
AZs consist of one or more discrete data centers, each with redundant power,
networking, and connectivity, housed in separate facilities.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 4-3 Cloud Platforms and Cloud Applications

These AZs offer you the ability to operate production applications and databases
that are more highly available, fault tolerant, and scalable than would be possible
from a single data center.
The AWS cloud operates 42 AZs within 16 geographic regions around the world,
with five more availability zones and two more regions coming online in 2017.
Each availability zone is designed as an independent failure zone. This means that
availability zones are physically separated within a typical metropolitan region
and are located in lower risk flood plains.

XEN Components
AWS consists of many cloud services that you can use in combinations tailored to
your business or organizational needs.
With Amazon Web Services you will find a complete cloud platform ready to use
for virtually any workload.
The user requests to the server by the method such as e-mail either to register or
to transfer the domain.
Your request which includes all information will be sent to Amazon API Gateway
restful service.
API Gateway will transfer the collected user information to an AWS Lambda
function.

AWS Lambda function will generate an e-mail and forward it to the 3rd party
mail server using Amazon SES.
Components of Amazon Web Service architecture are Amazon API Gateway, AWS
Lambda, Amazon Simple Email Service.

API Gateway is a front-door to access data, business logic and functionality. API
Gateway will provide a restful API endpoint for our AWS Lambda function.
API works at small as well as large-scale and helps developers to manage,
spectator, create and provide security to the API's.

Ji == - =—aib
E200) X= O

HTTP post LL
Contact US form requpst
"9 APl gateway |7 Amazon SES | 3rd party mail Business
submit request server owner

Fig. 4.1.1 AWS

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 4-4 Cloud Platforms and Cloud Applications

¢ AWS Lambda is a compute service that runs your back-end code and responds to
events such as object uploads to Amazon S3 bucket, Dynamo DB or in-app
activity. The Lambda function will get all the information from a user through API
Gateway.
e Amazon Simple email service helps us to send e-mail with minimal setup and
maximum deliverability. It is integrated with AWS management console so that
you can monitor your sending activity. Amazon Simple Email Service helps us by
monitoring insecurity.

EXE] Advantages and Disadvantages of AWS


Advantages :
1. Easy to use.
2. No capacity limits : Organizations launch different projects and the guess what
capacity they will need.
Provides speed and agility.
Secure and reliable : AWS provides security and also helps to protect the
privacy as it is stored in AWS data centers.

Disadvantages :
1. Limitations of Amazon EC2 : AWS sets default limits on resources which vary
from region to region. These resources consist of images, volumes, and
snapshots.
Technical support fee : AWS charges you for immediate support.
Security Limitations.

1. Write a note on services offered by Amazon. SPPU : Dec.-19, End Sem, Marks 8

4.2 | Elastic Cloud Computing SPPU : June-19, Dec.-19


e Amazon Elastic Compute Cloud (Amazon EC2) is a web service that provides
resizable compute capacity in the cloud. It is designed to make web-scale
computing easier for developers and system administrators.
e The Amazon EC2 simple web service interface allows you to obtain and configure
capacity with minimal friction. It provides you with complete control of your
computing resources and lets you run on Amazon's proven computing
environment.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 4-5 Cloud Platforms and Cloud Applications

¢ Amazon EC2 reduces the time required to obtain and boot new server instances
(called Amazon EC2 instances) to minutes, allowing you to quickly scale capacity,
both up and down, as your computing requirements change.
e Amazon EC2 changes the economics of computing by allowing you to pay only
for capacity that you actually use. Amazon EC2 provides developers and system
administrators the tools to build failure resilient applications and isolate
themselves from common failure scenarios.
e EC2 allows creating Virtual Machines (VM) on-demand. Pre-configured template
Amazon Machine Image (AMI) can be used get running immediately. Creating
and sharing your own AMI is also possible via the AWS marketplace.
e Amazon Machine Image (AMI) is a
template for software configuration
(Operating System, Application
Server, and Applications). Fig. 4.2.1 Host computer
shows AMI and instance. AMI
each instance computing power and memory is different
e Instance is a AMI running on
: : [ Instance |
virtual servers in the cloud. Each Launch instances (Cnstance )
instance type offers different of any type A i |
Inst
compute and memory facilities.
Create an Amazon Machine Image Host computer
(AMI) containing your applications, Fig. 42.1 AMI and instance
libraries, data and
associated configuration settings.Or use pre-configured, templated images to get
up and running immediately.
e Auto scaling allows automatically scale of the capacity up seamlessly during
demand spikes to maintain performance and scales down during demand lulls to
minimize costs. Auto Scaling increases resources when demand is high and reduces them when demand is low.

¢ Elastic load balancing automatically distributes incoming application traffic across


multiple Amazon EC2 instances. If provide tools to build failure resilient
applications by launching application instances in separate availability zones.
eo Pay only for resources actually consume, instance-hours. VM Import/Export
enables you to easily import virtual machine images from your existing
environment to Amazon EC2 instances and export them back at any time.
e Boto is a Python package that provides programmatic connectivity to Amazon
Web Services.
launching an EC2 instance :
#/usr/bin/python
import boto.ec2

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 4-6 Cloud Platforms and Cloud Applications

conn = boto.ec2.connect_to_region("us-west-2")
conn.run_instances(
'ami-6ac2a8ba’,
key_name='nitheesh_oregon’,
instance type='"1.micro,
security _groups=[nitheesh_oregon']
)
Stop instances :
#/usr/bin/python
import boto.ec2
conn = boto.ec2.connect_to_region('us-west-2")
conn.stop_instances(instance_ids=[instance-id-1,
'instance-id-2'])
e Boto supports more than fifty Amazon services, running the whole range from
compute, database, application and payments and billing.
e EC2 functions :
1. Load variety of operating system.
2. Install custom applications.
3. Manage network access permission.
4. Run image using as many/few systems as you desire.
e EC2 advantages :
1. Amazon EC2 enables you to increase or decrease capacity within minutes.
2. User have complete control of your Amazon EC2 instances.
3. Support flexible cloud hosting services
4 Secure : Amazon EC2 works in conjunction with Amazon VPC to provide
security and robust networking functionality.
5. Reliable : Amazon EC2 offers a highly reliable environment where replacement
instances can be rapidly and predictably commissioned.

(4.2.1 | Configuring Amazon EC2 Linux Instances

eo let's get started with Amazon Elastic Compute Cloud (Amazon EC2) by
launching, connecting to and using a Linux instance. An instance is a virtual
server in the AWS cloud. With Amazon EC2, you can setup and configure the
operating system and applications that run on your instance.
e¢ When you sign up for AWS, you can get started with Amazon EC2 using the
AWS Free Tier.
e The instance is an Amazon EBS-backed instance (meaning that the root volume is
an EBS volume). You can either specify the availability zone in which your

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 4-7 Cloud Platforms and Cloud Applications

instance runs or let Amazon EC2 select an availability zone for you. When you
launch your instance, you secure it by specifying a key pair and security group.
When you connect to your instance, you must specify the private key of the key
pair that you specified when launching your instance.
Various steps to configure Amazon EC2 Linux instance is shown in Fig. 4.2.2.

Step 1: Step 2: Step 3: Step 4 :


Signup for Create an Create a Create a virtual
AWS IAM user key pair private cloud

Step 8: Step 7: Step 6 : Step 5:


Cleanup Connect to Launch on Create a
instance instance instance security group

Fig. 4.2.2 Steps to signup for EC2


Step 1 : SignUp for AWS
When you signup for Amazon Web Services (AWS), your AWS account is
automatically signed up for all services in AWS, including Amazon EC2. You are
charged only for the services that you use.
With Amazon EC2, you pay only for what you use. If you are a new AWS
customer, you can get started with Amazon EC2 for free.
Step 2 : Create an IAM user
Services in AWS, such as Amazon EC2, require that you provide credentials when
you access them, so that the service can determine whether you have permission
to access its resources. The console requires your password.
You can create access keys for your AWS account to access the command line
interface or API. However, we don't recommend that you access AWS using the
credentials for your AWS account; we recommend that you use AWS Identity and
Access Management (IAM) instead.
Create an IAM user and then add the user to an JAM group with administrative
permissions or grant this user administrative permissions. You can then access
AWS using a special URL and the credentials for the JAM user. If you signed up
for AWS but have not created an IAM user for yourself, you can create one using
the IAM console.
Step 3 : Create a key pair
AWS uses public-key cryptography to secure the login information for your
instance. A Linux instance has no password; you use a key pair to log in to your
instance securely. You specify the name of the key pair when you launch your
instance, then provide the private key when you log in using SSH.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 4-8 Cloud Platforms and Cloud Applications

If you haven't created a key pair already, you can create one using the Amazon
EC2 console. Note that if you plan to launch instances in multiple regions, you'll
need to create a key pair in each region.
Step 4 : Create a Virtual Private Cloud (VPC)
Amazon VPC enables you to launch AWS resources into a virtual network that
you've defined, known as a Virtual Private Cloud (VPC). The newer EC2 instance
types require that you launch your instances in a VPC. If you have a default VPC,
you can skip this section and move to the next task, create a security group. To
determine whether you have a default VPC, open the Amazon EC2 console and
look for default VPC under account attributes on the dashboard.
Step 5 : Create a security group
Security groups act as a firewall for associated instances, controlling both inbound
and outbound traffic at the instance level. You must add rules to a security group
that enable you to connect to your instance from your IP address using SSH. You
can also add rules that allow inbound and outbound HTTP and HTTPS access
from anywhere. Note that if you plan to launch instances in multiple regions,
you'll need to create a security group in each region.
Step 6 : Launch an instance
You can launch a Linux instance using the AWS management console as described
in the following procedure.
1. Open the Amazon EC2 console at https://console.aws.amazon.com/ec2/.
2. From the console dashboard, choose Launch Instance.

3. The Choose an Amazon Machine Image (AMI) page displays a list of basic
configurations, called Amazon Machine Images (AMIs), that serve as templates
for your instance. Select an HVM version of Amazon Linux 2. Notice that these
AMIs are marked "Free tier eligible."
4. On the Choose an Instance Type page, you can select the hardware
configuration of your instance. Select the t2.micro type, which is selected by
default. Notice that this instance type is eligible for the free tier.
5. Choose Review and Launch to let the wizard complete the other configuration
settings for you.
6. On the Review Instance Launch page, under security groups, you'll see that
the wizard created and selected a security group for you. You can use this
security group or alternatively you can select the security group that you
created when getting setup using the following steps.
a) Choose Edit security groups.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 4-9 Cloud Platforms and Cloud Applications

b) On the Configure Security Group page, ensure that Select an existing


security group is selected.
c) Select your security group from the list of existing security groups and
then choose Review and Launch.
On the Review Instance Launch page, choose Launch.
When prompted for a key pair, select Choose an existing key pair, then select
the key pair that you created when getting setup. When you are ready, select
the acknowledgement check box and then choose launch instances.
A confirmation page lets you know that your instance is launching. Choose
View Instances to close the confirmation page and return to the console.
10. On the Instances screen, you can view the status of the launch. It takes a short
time for an instance to launch. When you launch an instance, its initial state is
pending. After the instance starts, its state changes to running and it receives a
public DNS name.
11. It can take a few minutes for the instance to be ready so that you can connect
to it. Check that your instance has passed its status checks; you can view this
information in the status checks column.

Step 7 : Connect to your Instance


Several ways to connect to your Linux instance is shown in Table 4.2.1.

Your computer OS Topic

Linux Connecting to your Linux instance using SSH.

Windows Connecting to your Linux instance from Windows using PuTTY.

Connecting to your Linux instance from Windows using Windows


Subsystem for Linux.
Other Connecting to your Linux instance using MindTerm
Table 4.2.1 Ways to connect to Linux instance

Step 8 : Cleanup your instance


e After you've finished with the instance, you should cleanup by terminating the
instance.
e Terminating an instance effectively deletes it; you can't reconnect to an instance
after you've terminated it.
e If you launched an instance that is not within the AWS free tier, youll stop
incurring charges for that instance as soon as the instance status changes to
shutting down or terminated. If you'd like to keep your instance for later, but not
incur charges, you can stop the instance now and then start it again later.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 4-10 Cloud Platforms and Cloud Applications

¢ To terminate your instance following steps can be used :


1) In the navigation pane, choose instances. In the list of instances, select the
instance.
2) Choose actions, instance state, terminate.

3) Choose yes, terminate when prompted for confirmation.


e Amazon EC2 shuts down and terminates your instance. After your instance is
terminated, it remains visible on the console for a short while and then the entry
is deleted.

(4.2.2 | Amazon S3

e Amazon 53 has a simple web services interface that you can use to store and
retrieve any amount of data, at any time, from anywhere on the web. S3 can serve
as a raw data store for IoT systems for storing raw data, such as sensor data, log
data, audio and video data.

AWS

Photo

; in Le rm
Response Queue Photo
prom)
[0]
ETE

, NEE =
> Web rN
OW J server Photo 2
oO
omQueue Photo w
processing
\ server /

Fig. 4.2.3 Amazon S3 working

e Features :
1. Unlimited storage
Highly scalable : In terms of storage, request rate and concurrent users.
won

Reliable : Store redundant data in multiple facilities and on multiple devices.


Secure : Flexibility to control who / how / when / where to access the data.

5. Performance : Choose region to optimize for latency / minimize costs.


e Example : Online photo processing service.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 4-11 Cloud Platforms and Cloud Applications

Procedure :
1. Web server receive request.
Put request message in the queue.
Pictures stored in S3.
SE

Multiple EC2 instances run photo processing.


Put back in the queue.
Return
e Store data on Amazon's distributed system containing multiple servers within
Amazon's data center locations. Amazon doesn't offer you a GUI based tool to
access your data. You can use one of the several tools online or build one through
APIs. to access the data
¢ Amazon EC2 provides three type of storage option : Amazon EBS, Amazon S3 and
Instance Storage. Amazon EBS (Elastic Block Store) provides with persistent, continuing
to do

block-level storage. Basically additional hard disk that you can attach to instance.
something

It suitable for apps which require database, filesystem, block level storage.
e A bucket is a container for objects stored in Amazon S3. Every object is contained
in a bucket. For example, if the object named "photos/puppy.jpg" is stored in the
rakshita bucket, then it is addressable using the URL
http: / /rakshita.s3.amazonaws.com /photos/pu
® Buckets serve several purposes : They organize the Amazon S3 namespace at the
highest level, they identify the account responsible for storage and data transfer
charges, they play a role in access control and they serve as the unit of
aggregation for usage reporting.
e Objects are the fundamental entities stored in Amazon S3. Objects consist of object
data and metadata. The data portion is opaque to Amazon S3. The metadata is a
set of name-value pairs that describe the object. These include some default
metadata, such as the date last modified and standard HITP metadata, such as
content-type. You can also specify custom metadata at the time the object is
stored.
e A key is the unique identifier for an object within a bucket. Every object in a
bucket has exactly one key. Because the combination of a bucket, key and version
ID uniquely identify each object, Amazon S3 can be thought of as a basic data
map between "bucket + key + version" and the object itself. Every object in
Amazon 53 can be uniquely addressed through the combination of the web service
endpoint, bucket name, key and optionally, a version.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 4-12 Cloud Platforms and Cloud Applications

Regions : You can choose the geographical region where Amazon S3 will store
the buckets you create. Objects stored in a region never leave the region unless
you explicitly transfer them to another region.

( N N\ NO)
Instance A Snapshot

= nd
\.
J
506
ephemeral0 ephemeral! ephemeral? ephemeral3
0]
\_ Instance store yp, \_ ) \_ )

Host computer Amazon EBS Amazon S83

Fig. 4.2.4 Amazon EBS and S3

(4.2.3 | Amazon CloudWatch

It is a monitoring service for AWS cloud resources and the applications user run
on AWS.
User can use Amazon CloudWatch to collect and track metrics, collect and
monitor log files and set alarms.
Amazon CloudWatch can monitor AWS resources such as Amazon EC2 instances,
Amazon DynamoDB tables and Amazon RDS DB Instances, as well as custom
metrics generated by user applications and services and any log files user
applications generate.
User can use Amazon CloudWatch to gain system-wide visibility into resource
utilization, application performance and operational health. User can use these
insights to react and keep user application running smoothly.
CloudWatch stores its data for two weeks, making it possible to track metrics
across an extended period.
CloudWatch is enabled when you create an account. Thereafter, you simply select
(or define) the metrics to track and then use the metrics that are generated as you
choose.
CloudWatch provides a lot of monitoring for free. For example, for any EC2
instances you have, you get ten metrics at 5-minute intervals with no cost.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 4-13 Cloud Platforms and Cloud Applications

1. Explain the steps for configuring a server for EC2. [Elgg UREI COCR TICT We ct gy
2. Explain the steps to create an Amazon S3 Bucket and managing associated objects ?
SPPU : Dec.-19, End Sem, Marks 9

3. Explain steps to configure server for EC2. SPPU : Dec.-19, End Sem, Marks 9

ZX) Amazon Storage System


e Amazon S3 defines a bucket name as a series of one or more labels, separated by
periods, that adhere to the following rules : The bucket name can be between 3
and 63 characters long, and can contain only lower-case characters, numbers,
periods, and dashes
e Amazon S3 defines a bucket name as a series of one or more labels, separated by
periods, that adhere to the following rules :
1. The bucket name can be between 3 and 63 characters long, and can contain
only lower-case characters, numbers, periods, and dashes.
Each label in the bucket name must start with a lowercase letter or number.

The bucket name cannot contain underscores, end with a dash, have
consecutive periods, or use dashes adjacent to periods.
4. The bucket name cannot be formatted as an IP address (198.51.100.24).
¢ A bucket is owned by the AWS account that created it. By default, you can create
up to 100 buckets in each of your AWS accounts. If you need additional buckets,
you can increase your bucket limit by submitting a service limit increase
¢ The following are the rules for naming S3 buckets in all AWS Regions :
Bucket names must be unique across all existing bucket names in Amazon S3.
Bucket names must comply with DNS naming conventions.
EA

Bucket names must be at least 3 and no more than 63 characters long.


Bucket names must not contain uppercase characters or underscores.
Bucket names must start with a lowercase letter or number.
SEU

Bucket names must be a series of one or more labels. Adjacent labels are
separated by a single period (.). Bucket names can contain lowercase letters,
numbers, and hyphens. Each label must start and end with a lowercase letter
or a number.
Bucket names must not be formatted as an IP address (for example,
192.168.5.4).

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 4-14 Cloud Platforms and Cloud Applications

8. When you use virtual hosted-style buckets with Secure Sockets Layer (SSL), the
SSL wildcard certificate only matches buckets that don't contain periods. To
work around this, use HTTP or write your own certificate verification logic.
We recommend that you do not use periods (".") in bucket names when using
virtual hosted-style buckets.

mm Amazon Database Services


Amazon Relational Database Service (RDS) is a web service that makes it easy to
set up, operate and scale a relational database in the cloud. RDS gives access to
the capabilities of a familiar MySQL, Oracle or Microsoft SQL Server database
engine. Code, applications and tools already used with existing databases can be
used with RDS.
Amazon RDS is just a replacement to running your own database server. With a
simple sequence of commands at the console, you can choose from two
commercial DBMS (Oracle or Microsoft SQL Server) or two open source DBMS
(MySQL or PostgreSQL).
Amazon RDS automatically patches the database software and backs up the
database, storing the backups for a user-defined retention period and enabling
point-in-time recovery.
Amazon RDS provides scaling the compute resources or storage capacity
associated with the Database Instance.Pay only for the resources actually
consumed, based on the DB Instance hours consumed, database storage, backup
storage and data transfer.
On-Demand DB Instances let you pay for compute capacity by the hour with no
long-term commitments. Reserved DB Instances give the option to make a low,
one-time payment for each DB Instance and in turn receive a significant discount
on the hourly usage charge for that DB Instance.
Amazon RDS provides you six familiar database engines to choose from, including
Amazon Aurora, PostgreSQL, MySQL, MariaDB, Oracle and Microsoft SQL Server.
Create the actual RDS instance :
instance = conn.create_dbinstance ( id=identifier, instance_class="db." + instance class,
allocated_storage=allocated_storage,
engine=engine, db_name=db_ name, master_username=master_username,
master_password=master password,
db_subnet group name=identifier, vpc_security _groups=security_ groups, port=port)

Advantages :
1. Fast and easy to administer : Amazon RDS makes it easy to go from project
conception to deployment.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 4-15 Cloud Platforms and Cloud Applications

2, Available and durable : Amazon RDS runs on the same highly reliable
infrastructure used by other Amazon Web Services.
3. Highly scalable : You can scale your database's compute and storage
resources.
4. Secure : Amazon RDS makes it easy to control network access to your
database.
also from ppt
XX Dynamo DB
DynamoDB is a fast, fully managed NoSQL database service that makes it simple
and cost-effective to store and retrieve any amount of data and serve any level of
request traffic.
All data items are stored on Solid State Drives (SSDs) and are replicated across
3 Availability Zones for high availability and durability.
DynamoDB tables do not have fixed schemas and each item may have a different
number of attributes.
DynamoDB has no upfront costs and implements a pay as you go plan as a. a
flat hourly rate based on the capacity reserved.
With DynamoDB, you can create database tables that can store and retrieve any
amount of data and serve any level of request traffic. You can scale up or scale
down your tables’ throughput capacity without downtime or performance
degradation and use the AWS management console to monitor resource utilization
and performance metrics.
DynamoDB allows you to delete expired items from tables automatically to help
you reduce storage usage and the cost of storing data that is no longer relevant.
The first step in accessing DynamoDB is to create a connection to the service :
import boto.dynamodb
conn = boto.dynamodb.connect_to_region( 'us-west-2,
aws _access_key_id='<YOUR_AWS KEY ID>',
aws _secret_access_key='<YOUR_AWS_SECRET KEY>')
Python program for creating a DynamoDB table with other operation :
import boto.dynamodb
from boto.dynamodb.condition import *
connection = boto.dynamodb.connect_to_region('eu-west-1')
table = connection.get_table(‘table')

id ="1'
timestamp = 1234
attrs = {
'key1" ‘'valuel!,
'key2" set(['value2', 'value3'])
}
TECHNICAL PUBLICATIONS® - an up-thrust for knowledge
Cloud Computing 4-16 Cloud Platforms and Cloud Applications

# create
item = table.new_item(hash key=id, range key=timestamp, attrs=attrs)
item.put()

# read
item = table.get_item(hash key=id)
key2 = list(item[key2'])

# update
item[’key1'] = 'foo'
item['key3'] = 'bar'
item.put()

# query
table.query(hash_key=id, range_key_condition=LT(1500))

# scan
table.scan(scan_filter={'key1": EQ('foo')})

# delete
item = table.get_item(hash_key=id)
item.delete()

Advantages :
1. Flexible : Amazon DynamoDB supports both document and key-value data
structures.
2. Fully managed : Amazon DynamoDB is a fully managed cloud NoSQL
database service.
3. Highly scalable : When you create a table, simply specify how much request
capacity you require.
4. Event-driven programming : Amazon DynamoDB integrates with AWS lambda
to provide triggers

XX] Difference between DynamoDB and Amazon S3


DynamoDB Amazon S3

DynamoDB is database $3 is file storage


Used in semi structured data It is used for unstructured data

Size limit is 400 kb Size limit is 5 TB


6. A NoSQL document-oriented database. An object storage service.
7. Data is stored in tables as items. Data is stored in buckets as objects.
TECHNICAL PUBLICATIONS® - an up-thrust for knowledge
8. DynamoDB does not support built-in versioning, but you S3 supports automated versioning to record and recover multiple
can implement it using atomic counters. versions of your objects.
Cloud Computing 4-17 Cloud Platforms and Cloud Applications

DynamoDb supports two kinds of primary $3 uses unique Ids called Keys to retrieve files
keys, Partition Key and Partition key and Sort from the bucket
key.
DynamoDb is used to store key-value. It uses S3 stores files in a flat organisation of
items and attributes for its tables containers called Buckets

Microsoft Cloud Services : Azure

Windows Azure is a cloud computing platform and infrastructure, created by


Microsoft, for building, deploying and managing applications and services through
a global network of Microsoft - managed data centers.
Azure queue storage is a service for storing large numbers of messages that can be
accessed from anywhere in the world via authenticated calls using HTTP or
HTTPS. A single queue message can be up to 64 KB in size, and a queue can
contain millions of messages, up to the total capacity limit of a storage account.
Azure is a virtualized infrastructure to which a set of additional enterprise services
has been layered on top, including, a virtualization service called Azure AppFabric
that creates an application hosting environment. AppFabric is a cloud-enabled
version of the NET framework.
Windows Azure is Microsoft's application platform for the public Cloud.
Applications can be deployed on to Azure in various models
Windows Azure is used to :
1. Build a web application that runs and stores its data in Microsoft data centers.
2. Store data while the applications that consume this data run on premise
(outside the public Cloud).
3. Create virtual machines to develop and test, or run SharePoint and other
out-of-the-box applications.
4. Develop massively scalable applications with many users.
5. Offer a wide range of services
Azure has three components : compute, storage and fabric
1. Compute : Windows Azure provides a hosting environment for managed code.
It provides a computation service through roles. Windows Azure supports
three types of roles :
a) Web roles used for web application programming and supported by IIS7.
b) Worker roles are also used for background processing of web roles.
c) Virtual Machine (VM) roles are generally used for migrating windows
server applications to Windows Azure in an easy way.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 4-18 Cloud Platforms and Cloud Applications

2. Storage : Windows Azure provides storage in the cloud. It provides four


different types of storage services :
a) Queues for messaging between web roles and worker roles.
b) Tables for storing structural data.
c) BLOBs (Binary Large Objects) to store text, files or large data.
d) Windows Azure Drives (VHD) to mount a page blob. They can easily be
downloaded and uploaded via blobs.
3. AppFabric provides infrastructure services for developing, deploying and
managing Windows Azure application. It provides five services: Service bus,
Access, Caching, Integration and Composite.
e Fig. 4.5.1 shows Windows Azure platform architecture.

Cloud applications

« 3
0 o113
cw Li Net = 0
0 + SQL Azure ve Q Z
=
© QO
services services 2
+=
[0]0h
QE . S
54a Cloud platform services g 2
a rs
oO [0]
5
<C D o 5
[0]
° Windows Azure =
=
Compute Storage Network
resources resources resources

Fabric

Hosting platform

Windows Azure platform

Fig. 4.5.1 Windows Azure platform architecture

e Microsoft Azure is a cloud computing service created by Microsoft for building,


testing, deploying and managing applications and services through a global
network of Microsoft-managed data centers.
e It provides software as a service (SaaS), platform as a service and infrastructure as
a service and supports many different programming languages, tools and
frameworks, including both Microsoft-specific and third-party software and
systems.
¢ Windows Azure provides resources and services for consumers. For example,
hardware is abstracted and exposed as compute resources.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 4-19 Cloud Platforms and Cloud Applications

Physical storage is abstracted as storage resources and exposed through very


well-defined interfaces.
A common windows fabric abstracts the hardware and the software and exposes
virtual compute and storage resources.
Each instance of an application is automatically managed and monitored for
availability and scalability.
If an application goes down, the Fabric is notified and a new instance of the
application is created. Because virtualization is a key element in cloud computing,
no assumption must be made on the state of the underlying hardware hosting the
application.
Advantages of Microsoft Azure
1. Microsoft Azure offers high availability
2. It offers you a strong security profile
3. It is a cost-effective solution for an IT budget.
4. Azure allows you to use any framework, language, or tool.
5. Azure allows businesses to build a hybrid infrastructure.

4.6 | Cloud Computing Applications


Cloud Adoption is a strategic move by organisations of reducing cost, mitigating
risk and achieving scalability of data base capabilities. Cloud adoption may be up
to various degrees in an organisation, depending on the depth of adoption. In fact
the depth of adoption yields insight into the maturity of best practices,
enterprise-ready cloud services availability.
A variety of industries benefit from cloud adoption, including healthcare,
marketing and advertising, retail, finance and education. Benefits include :

Healthcare : Fueled by digital and social consumer behaviors and the need for
secure and accessible Electronic Health Records (EHRs), hospitals, clinics and other
medical organizations are using cloud computing for document storage, marketing
and human resources.
Marketing and Advertising : In an industry dependent on social media, as well as
the quick creation and publishing of customer - relevant content, agencies are
using hybrid cloud adoption strategies to deliver critical client messages to their
local and worldwide audiences.
Retail : A successful e-commerce strategy requires a sound Internet strategy. With
the help of cloud adoption, internet-based retail is able to effectively market to
customers and save their product data for less money.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 4-20 Cloud Platforms and Cloud Applications

e Finance : Efficient expense management, human resources and customer


communications are three of the most important business needs of today's finance
organizations. For these reasons, financial services institutions are now placing
their email platforms and marketing tools in the cloud.

EXXN Healthcare : ECG Analysis in the Cloud


o Electrocardiograph (ECG) analysis brings a lot of technical concerns because ECG
is one of the tools frequently used in the diagnosis of cardiovascular disease.
e In the cloud system, medical data can be gathered and distributed automatically to
medical practitioners anywhere in the world. From there, doctors in the field have
the capability of returning input to specific patients.
e Cloud computing technologies allows the remote monitoring of a patient's heart
beat data. Through this way the patient at risk can be constantly monitored
without going to the hospital for ECG analysis. At the same time the Doctor's can
instantly be notified with cases that need's their attention. Fig. 4.6.1 shows cloud
ECG process.

— Storage Ia
Wireless ECG sensor - Ny

Embedded bluetooth
enabled
Data processor and
communication module

ECG data
ECG sensor module SaaS analysis
software

Of .
Bluetooth connectivity

®
Wireless/Mobile 3G network

Vv |,
Windows Od
phone

Fig. 4.6.1 Cloud ECG process

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 4-21 Cloud Platforms and Cloud Applications

¢ The respective information is transmitted to the patient's mobile device that will
immediately forwarded to the cloud- hosted web services for analysis. The entire
web services from the front end of a platform that is completely hosted in the
cloud that consist of three layers : Saas, Paas, laas.

EX:X] Biology : Protein Structure Prediction


e The prediction of the protein structure is the inference from its amino acid
sequence of the three - dimensional structure of a protein, that is, the prediction of
its secondary and tertiary structure from the primary structure.
¢ Fig. 4.6.2 shows protein structure prediction using cloud.

PredictionManager
User Input queue
role

Output queue

DB with - oo
prediction results _ ~~ Prediction Prediction
- input output
queue

3D protein
Azure structures
(PDB files)

Prediction\WWorker roles

Microsoft azure public cloud

Fig. 4.6.2
e The Predict Protein cloud solution builds upon the open source operating system
Debian and provides its functionality as a set of free software packages. Bio-Linux
is an operating system for bioinformatics and computational biology.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 4-22 Cloud Platforms and Cloud Applications

¢ The latest Bio-Linux release provides more than 500 bioinformatics programs on
an Ubuntu Linux base. Ubuntu is a "derivative" operating system based on
Debian, with its own additions.

e Cloud BioLinux is a comprehensive cloud solution that is derived from Bio-Linux


and Ubuntu. Debian derivatives can easily share packages between each other. For
example, Debian packages are automatically incorporated in Ubuntu and are also
usable in Cloud BioLinux.

EXE] Geosciences : Satellite Image Processing


e Satellite image processing plays a vital role for research and developments in
Astronomy, Remote Sensing, GIS, Agriculture Monitoring, Disaster Management
and many other fields of study. Satellite images are recorded in digital forms and
then processed by the computers to extract information.
e Satellite remote sensing generates hundreds of giga - bytes of raw images that
need to be further processed to become the basis of several different GIS products.
Fig. 4.6.3 shows cloud environment for satellite data processing.

Portal (SaaS)

Distribution

Local storage

Private cloud

Ng cloud

Fig. 4.6.3 Cloud environment for satellite data processing

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 4-23 Cloud Platforms and Cloud Applications

This process requires both I/O and compute-intensive tasks. Large images need to
be moved from a ground station's local storage to compute facilities, where several
transformations and corrections are applied.
Cloud computing provides the appropriate infrastructure to support such
application scenarios.
Satellite data processing consists of various methods to correct the radiometric
errors and geometric distortions in the basic data generated by the sensor; this
data is termed as Level-0.
The procedures like georeferencing and registration applied on the Level-0 data to
generate the products such as;
1. Level 1 - Radio metrically corrected and geometrically corrected only for earth
rotation (Browse product)
2. Level 2 - Both radiometric and geometrically corrected (Standard product)

4.6.4 | Business and Consumer Applications : CRM and ERP

Customer Relationship Management (CRM) is a strategy for managing all your


company's relationships and interactions with your customers and potential
customers. It helps you improve your profitability.
Cloud ERP is Software as a Service that allows users to access Enterprise Resource
Planning (ERP) software over the Internet. Cloud ERP generally has much lower
upfront costs, because computing resources are leased by the month rather than
purchased outright and maintained on premises. Cloud ERP also gives companies
access to their business-critical applications at any time from any location.
While technically the only difference between Cloud ERP and on-premises ERP is
where the software is physically located, there are other significant differences.
Here we explain some of the key characteristics and advantages of Cloud ERP
software.
The Cloud is particularly valuable to Small and Medium - Size Businesses (SMB's)
because it provides access to full-function applications at a reasonable price
without a substantial upfront expenditure for hardware and software. Using the
right cloud provider, a company can rapidly scale their business productivity
software as their business grows or a new company is added.
Cloud ERP has been proven to reduce costs in many ways because it :
a) Avoids upfront costs for all computing infrastructure such as hardware and
data servers.
b) Reduces IT support services because IT support is provided by the data center.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 4-24 Cloud Platforms and Cloud Applications

c) Eliminates paying upfront for application software licenses in favor of a


monthly fee.
d) Shrinks the cost of maintaining and supporting those applications since the
cloud vendor handles the updates and upgrades.

4.6.5 | Social Networking

Social Network Analysis (SNA) is an important and valuable tool for knowledge
extraction from massive and un-structured data. Social network provides a
powerful abstraction of the structure and dynamics of diverse kinds of
inter-personal connection and interaction.
Facebook is a social networking service and website that connects people with
other people and share data between people. A user can create a personal profile,
add other users as friends, exchange data, create and join common interest
communities.
Twitter is a social net-working and microblogging service. The users of Twitter can
exchange text-based posts called tweets. A tweet is a maximum 140 characters
long but can be augmented by pictures or audio recording. The main concept of
Twitter was to build a social network formed by friends and fol-lowers. Friends
are people who you follow, followers are those who follow you.
The role of social networks in labor markets deserves attention for at least two
reasons : First, because of the central role networks play in disseminating
information about job openings they place a critical role in determining whether
labor markets function efficiently and second, because network structure ends up
having implications for things like human capital investment as well as inequality.
Social Network Analysis (SNA) primarily focuses on applying analytic techniques
to the relationships between individuals and groups and investigating how those
relationships can be used to infer additional information about the individuals and
groups.
SNA is used in a variety of domains. For example, business consultants use SNA
to identify the effective relationships between workers that enable work to get
done; these relationships often differ from connections seen in an organizational
chart.
Law enforcement personnel have used social networks to analyze terrorist
networks and criminal networks. The capture of Saddam Hussein was facilitated
by social network analysis : Military officials constructed a network containing
Hussein's tribal and family links, allowing them to focus on individuals who had
close ties to Hussein

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 4-25 Cloud Platforms and Cloud Applications

¢ Collecting social network data used to be a tedious, labor - intensive process. In


fact, several notable dissertations came out of the researcher's being at the right
place and the right time to be able to observe a social conflagration and gather
data on it.
e Social network data collection is, by nature, more invasive and harder to
anonymize; survey instruments had to be approved by Institutional Review Boards
(IRBs) and administration of the surveys was tedious manual labor.
¢ Some of key challenges in this kind of data collection are :
1 Network boundaries are difficult to define.
2 People do not easily recall their network members and need appropriate
"prompts" to elicit them. In addition, networks are very large in general and
different social network members may have different importance depending
on the phenomenon studied.
3 Information about the network members needs to balance detail and
interviewee's burden.
e Most social network data collection can be divided into "whole" and "egocentric"
networks. Whole network studies examine actors "that are regarded for analytical
purposes as bounded social collectives”; actors in these studies are named in
closed lists, usually pre-defined and known a priori.
e Since these boundaries are very difficult to define in urban settings with large
populations, whole network studies are unpractical, making egocentric data
collection the only feasible method.
e Egocentric network studies concentrate in specific actors or egos and those who
have relations with them, called alters. That is, from the participant's perspective,
egocentric networks constitute a "network of me" or a network of actors with
whom the participant has some relationship.
e Egocentric network data is thus composed by two levels :
i) An ego-network level, constituted by the ego's characteristics and overall
network features; and
ii) An ego-alter level, constituted by the characteristics of each alter and alter-ego
ties.

Google Cloud Application : Google App Engine

¢ Google App Engine (GAE) is a Platform as a Service cloud computing platform for
developing and hosting web applications in Google-managed data centers.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 4-26 Cloud Platforms and Cloud Applications

* Google App Engine is a way to write your own web applications and have them
hosted on Google servers. It enables developers to build their web applications on
the same scalable system that power Google applications.
e An app is a piece of software which can run on the computer, internet, phone or
any other electronic device. Google refers to their online services as Apps. They
also sell a specific suite of services known as Google Apps.
* Google's providing both SaaS and PaaS solutions in cloud computing. Some of the
examples for SaaS solutions including Google Apps which including Gmail, Doc,
etc. and PaaS includes Google App engine.
e Services provided by App engine includes :
a) Platform as a Service (PaaS) to build and deploy scalable applications
b) Hosting facility in fully-managed data centers
c¢) A fully-managed, flexible environment platform for managing application
server and infrastructure.
d) Support in the form of popular development languages and developer tools.
¢ Major feature of Google App Engine :
1. Automatic scaling and load balancing.
Authentication using Google Accounts APL
ISU

Provides dynamic web services based on common standards.


Integration with other Google Cloud Services and APL
Support persistent storage, with query access sorting and transaction
management features.
e Google App engine offers users the ability to build and host web applications on
Google's infrastructure.

Google app
engine

Datastore Caching Authentication Mail

LN Memcache Google apps Gmail

Fig. 4.7.1

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 4-27 Cloud Platforms and Cloud Applications

¢ The App Engine offers a number of services that enable you to perform several
common operations when managing your application. The following APIs are
available to access these services :
1. Mail : Using the mail API, the developers can send email messages.
2. Memcache : The Memcache service gives the users the benefit of working
efficiently by providing high retrieval speed, even when multiple users access
the same application at the same instance of time.
3. Image manipulation : The Image service allows you to manipulate images of
your application. With the use of this API, you can resize, crop, rotate and flip
images in JPEG and PNG formats.
¢ In the PaaS space Google is a key player. App Engine is a platform to create, store
and run applications on Google's servers using development languages as java and
python.
¢ App Engine includes tools for managing the data store, monitoring the site and its
resource consumption and debugging and logging. A user can serve the app from
his own domain name using Google Apps.
¢ Key features of GAE programming mode using java and python.
e The Google App engine Software Development Kit (SDK) provides Java and
Python programming languages.
¢ The languages have their own web server application that contains all Google App
Engine services on a local computer. The web server also simulates a secure
sandbox environment.
e The Google App engine SDK has APIs and libraries including the tools to upload
applications. The architecture defines the structure of applications that run on the
Google App engine.
1. Python :
e The Google App engine allows implementation of applications using python
programming language and running them on its interpreter.
e The Google App engine provides rich APIs and tools for designing web
applications, data modeling, managing, accessing apps data, support for mature
libraries and frameworks like Django.
e The main characteristics of Google App engine are its DataStore, configuration file
app.yaml and how it serves an application.
2. Java :
e The Google App engine provides tools and APIs required for the development of
web applications that run on the Google App engine Java run time.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 4-28 Cloud Platforms and Cloud Applications

eo The application interacts with the environment using servlets and web
technologies like Java Server Pages (JSPs) which can be developed using Javaé.
e The GAE environment uses Java SE Runtime JRE platform 6 and libraries which
the applications can access using APIs.
e Java SDK has implementations for Java Data Objects (JDO) and Java Persistence
(JPA) interface.
e To exchange email messages with Google App engine, it provides the Google App
Engine mail service through the Java Mail APL
e Support for other languages like JavaScript, Ruby or Scalar is also provided by
Google App engine with the use of JVM compatible compilers and interpreters.
e When Google App engine gets a web request that corresponds to the URL
mentioned in the applications deployment descriptor it invokes a servlet
corresponding to the request and uses Java Servlets API to provide requested data
and accepts response data.
e Google App engine makes it easy to build an applications that runs reliably, even
under heavy load and with large amounts of data.
e App engine includes the below features :
a) Dynamic web serving, with full support for common web technologies.
b) Persistent storage with queries, sorting and transactions.
¢) Automatic scaling and load balancing.
d) APIs for authenticating users and sending email using Google accounts.
e) Scheduled tasks for triggering events at specified times and regular intervals.

ZX] Overview of OpenStack Architecture


e OpenStack is a recently open-sourced, IaaS cloud-computing platform founded by
Rackspace Hosting and NASA, and is used widely in industry
e OpenStack is an open-source cloud platform. OpenStack software controls large
pools of compute, storage, and networking resources throughout a data center, all
managed by a dashboard that gives administrators control while empowering their
users to provision resources through a web interface.
e To produce the ubiquitous Open-Source cloud computing platform that will meet
the needs of public and private cloud providers regardless of size, by being simple
to implement and massively scalable.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 4-29 Cloud Platforms and Cloud Applications

¢ Components of OpenStack are as follows :


1. Horizon - Dashboard : It provides a modular web-based user interface for all
the OpenStack services. With this web GUI, user can perform most operations
on your cloud like launching an instance, assigning IP addresses and setting
access controls.
2. Keystone is a framework for authentication and authorization for all the
OpenStack services. It handles API requests as well as providing configurable
catalog, policy, token and identity services. Keystone is a framework for
authentication and authorization for all the OpenStack services.
3. Nova : It provides virtual servers upon demand. Nova is the most complicated
and distributed component of OpenStack. A large number of processes
cooperate to turn end user API requests into running virtual machines.
4. Glance - Image Store : It provides discovery, registration and delivery services
for disk and server images.
5. Quantum - Network : It provides " network connectivity as a service " between
interface devices managed by other OpenStack services. The service works by
allowing users to create their own networks and then attach interfaces to them.
Quantum has a pluggable architecture to support many popular networking
vendors and technologies.
6. Cinder allows block devices to be exposed and connected to compute instances
for expanded storage & better performance.
7. Object store allows you to store or retrieve files. It provides a fully distributed,
API-accessible storage platform that can be integrated directly into applications
or used for backup, archiving and data retention.

EX] Multiple Choice Questions


Q.1 Amazon EC2 is a computing service, whereas Amazon SQS and Amazon S3 are
services.

a| cloud b| support
c| storage d| platform

Q.2 Google Cloud Storage is a RESTful online storage web service for
storing and accessing one's data on Google's infrastructure.

a| data b| information
c| block d| file

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 4-30 Cloud Platforms and Cloud Applications

Q.3 Google Cloud Print is a service that extends the printer's function to any device
that can connect to the

[al Intranet [b] Internet

WAN [d| LAN


Q.4 The AWS global infrastructure consists of multiple geographical locations which are
called

[a] division [b] block

regions [d] cluster

Q.5 Amazon Elastic Block Store provides persistent storage volumes for use
with Amazon EC2 instances in the AWS Cloud.

a] file [b] block


directory [d] all of these

Q.6 Amazon is a low-cost cloud storage service for data with longer
retrieval times offered by Amazon Web Services.

[a] Cloudnet [b] RedShift


DynamoDB [d] Glacier

Q.7 Amazon EC2 provides virtual computing environment, known as

[a] chunks [b] instances


block [d| messages

Q8 What does S3 stand for ?

[a] Super Storage Service [b] Storage Simple Service


Single Storage Service [d| Simple Storage Service

Q.9 Amazon S3 is a

[a] Content Delivery Network [b] Domain System


Key-Based object store [d] Relational database

Q.10 Which of these services are used to distribute content to end users using a global
network of edge locations ?

[a] CloudFront [b] Route 53


Virtual Private Cloud [d] CloudWatch

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 4-31 Cloud Platforms and Cloud Applications

Q.11 DynamoDB is a fast, fully managed database service that makes it simple
and cost-effective to store and retrieve any amount of data and serve any level of
request traffic.

[a] PL/SQL [b| NosQL


SQL [d| All of these
Q.12 S3 stores files in a flat organisation of containers called

la] chunks [b] block


Buckets [d] none of these

Q.13 Microsoft offers its own online collaboration tool called

[a] Azer [b] salespoint


Sharepoint [d] Aneka

Answer Keys for Multiple Choice Questions

Q1 b Q.2 d Q.3 b Q4 c

Q.5 b Q.6 d Q.7 b Q.8 d

Q.9 c Q.10 a Q.11 b 0.12 c

Q13 | «©

aaa

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 4-32 Cloud Platforms and Cloud Applications

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


UNIT -V

Security in Cloud Computing

Syllabus
Risks in Cloud Computing : Risk Management, Enterprise-Wide Risk Management, Types of Risks
in Cloud Computing. Data Security in Cloud : Security Issues, Challenges, advantages,
Disadvantages, Cloud Digital persona and Data security, Content Level Security.
Cloud Security Services : Confidentiality, Integrity and Availability, Security Authorization
Challenges in the Cloud, Secure Cloud Software Requirements, Secure Cloud Software Testing.

Contents
5.1 Risks in Cloud Computing
5.2 Enterprise-wide Risk Management
5.3 Types of Risks in Cloud Computing
54 Data Security in Cloud : Security Issues
and Challenges ~~ .....c.cviiinan.. March-20, -------------- Marks 5
5.5 Content Level Security
5.6 Cloud Security Services
5.7 Security Authorization Challenges in the Cloud
5.8 Secure Cloud Software Requirements
59 Secure Cloud Software Testing
5.10 Multiple Choice Questions

(5-1)
Cloud Computing 5-2 Security in Cloud Computing

5.1 | Risks in Cloud Computing


e Cloud security is the protection of data stored online via cloud computing
platforms from theft, leakage, and deletion. Methods of providing cloud security
include firewalls, penetration testing, tokenization, Virtual Private Networks
(VPN), and avoiding public internet connections.
¢ Cloud security refers to an array of policies, technological procedures, services,
and solutions designed to support safe functionality when building, deploying,
and managing cloud-based applications and associated data.
eo Cloud security is designed to protect the following, regardless of your
responsibilities :
a) Physical networks - Routers, electrical power, cabling, climate controls, etc.
b) Data storage - Hard drives, etc.

c) Data servers - Core network computing hardware and software


d) Computer virtualization frameworks - Virtual machine software, host
machines, and guest machines
e) Operating Systems (OS) - Software that houses
f) Middleware - Application Programming Interface (API) management,
g) Runtime environments - Execution and upkeep of a running program
h) Data - All the information stored, modified, and accessed

i) Applications - Traditional software services (email, tax software, productivity


suites, etc.)

j) End-user hardware - Computers, mobile devices, Internet of Things (IoT)


devices, etc.
¢ Cloud computing security addresses both physical and logical security issues
across all the different service models of software, platform and infrastructure. It
also addresses how these services are delivered in the Public, Private, Hybrid and
Community delivery models.

EXEN Risk Management


¢ Risk management is the process that allows business managers to balance
operational and economic costs of protective measures and achieve gains in
mission capability by protecting business processes that support the business
objectives or mission of the enterprise.
e Risk management is the total process used to identify, control and minimize the
impact of uncertain events. The objective of the risk management program is to

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 5-3 Security in Cloud Computing

reduce the risk of performing some activity or function to an acceptable level and
obtain senior management approval.
o Threat is a potential cause of an incident that may result in harm to a system or
organization.
e Vulnerability is a weakness of an asset (resource) or a group of assets that can be
exploited by one or more threats.
¢ Risk is potential for loss, damage, or destruction of an asset as a result of a threat
exploiting a vulnerability.
e Risk control is an important part of risk management. It involves determining
what to do with uncontrolled risks.
e Some questions to ask when selecting a risk control strategy are, "What is an
acceptable level of risk ?" and "What should I do about the risks ?"
e Risk control is often achieved by applying safeguards. Safeguard is anything that
removes a vulnerability or protects against one or more specific threats.
e Security risk analysis, otherwise known as risk assessment, is fundamental to the
security of any organization. It is essential in ensuring that controls and
expenditure are fully commensurate with the risks to which the organization is
exposed.
¢ However, many conventional methods for performing security risk analysis are
becoming more and more untenable in terms of usability, flexibility and critically...
in terms of what they produce for the user.
e Security in any system should be commensurate with its risks. However, the
process to determine which security controls are appropriate and cost effective, is
quite often a complex and sometimes a subjective matter. One of the prime
functions of security risk analysis is to put this process onto a more objective
basis.
e There are a number of distinct approaches to risk analysis. However, these
essentially break down into two types : Quantitative and qualitative.

5.2 | Enterprise-wide Risk Management


e Enterprise Risk Management (ERM) is the overall management of risk for an
organization. As with governance, the contract defines the roles and
responsibilities for risk management between a cloud provider and a cloud
customer. And, as with governance, you can never outsource your overall
responsibility and accountability for risk management to an external provider.
¢ Risk management in cloud is based on the shared responsibilities model.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 5-4 Security in Cloud Computing

¢ Risk management process is as follows :


1. Define objective
2. Identify risk
3. Evaluate risk
4. Options and assortment of risk
5. Decision about implementation
6. Evolution and review
e Fig. 5.2.1 shows six step risk administration process.

Define Object
Identify risk

Evaluate risk

Options and assortment of risk

Decision about implementation


Evolution and review
Fig. 5.2.1 Six step risk administration process

Parameters Remarks

Define object o Administration program


Identify risk o Checklist, flowchart

o Inspections

e Internal records

Evaluate risk Significant or insignificant risk

Options and assortment of risk e How to deal with risk


Decision about implementation e Methods

Risks remedy
Evolution and review ¢ Risk administration

[EX] Types of Risks in Cloud Computing


¢ Risks in cloud computing is divided into internal and external.
e Loss of data : Data stored on cloud servers can be lost through a natural disaster,
malicious attacks, or a data wipe by the service provider.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 5-5 Security in Cloud Computing

e Increased customer agitation : A growing number of cloud service critics are keen
to see which service providers have weak security protocols and encourage
customers to avoid them.
e Attacks to deny service to legitimate users.
¢ Shared vulnerabilities : Cloud security is the responsibility of all concerned parties
in a business agreement.
e Contract breaches with clients and/or business partners : Contracts restrict how
business partners or clients use data and also who has the authorization to access
it.
e Malware attacks : Cloud services can be a vector for data exfiltration.
e Compliance violations : Organizations can quickly go into a state of
non-compliance, which puts them in the risk of serious repercussions
¢ top threats identified by Cloud Security Alliance (CSA) of cloud computing are as
follows :
1. Insecure Interfaces and APIs : Cloud Computing providers expose a set of
software interfaces or APIs that customers use to manage and interact with
cloud services. Provisioning, management, orchestration and monitoring are all
performed using these interfaces.
¢ Remediation : Analyze the security model of cloud provider interfaces.
2. Malicious insiders : The threat of a malicious insider is well-known to most
organizations. This threat is amplified for consumers of cloud services by the
convergence of IT services and customers under a single management domain,
combined with a general lack of transparency into provider process and
procedure.
¢ Remediation : Determine security breach notification processes.
3. Shared technology issues : IaaS vendors deliver their services in a scalable way
by sharing infrastructure.
e Remediation : Implement security best practices for installation/configuration.
Monitor environment for unauthorized changes/activity.
4. Data loss or leakage : There are many ways to compromise data. Deletion or
alteration of records without a backup of the original content is an obvious
example. Unlinking a record from a larger context may render it
unrecoverable, as can storage on unreliable media. Loss of an encoding key
may result in effective destruction.
¢ Remediation : Implement strong API access control. Encrypt and protect integrity
of data in transit.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 5-6 Security in Cloud Computing

5. Account or service hijacking : Attack methods such as phishing, fraud and


exploitation of software vulnerabilities still achieve results. Credentials and
passwords are often reused, which amplifies the impact of such attacks.
¢ Remediation : Prohibit the sharing of account credentials between users and
services. Leverage strong two-factor authentication techniques where possible.
Employ proactive monitoring to detect unauthorized activity.

5.4 | Data Security in Cloud : Security Issues and Challenges

e Cloud computing security challenges fall into three broad categories :


1. Data protection : Securing your data both at rest and in transit
2. User authentication : Limiting access to data and monitoring who accesses the
data
3. Disaster and data breach : Contingency planning
e Data protection : Data needs to be encrypted at all times, with clearly defined
roles when it comes to who will be managing the encryption keys.
e User authentication : Data resting in the cloud needs to be accessible only by those
authorized to do so, making it critical to both restrict and monitor who will be
accessing the company's data through the cloud. In order to ensure the integrity
of user authentication, companies need to be able to view data access logs and
audit trails to verify that only authorized users are accessing the data.
e Contingency planning : With the cloud serving as a single centralized repository
for a company's mission-critical data, the risks of having that data compromised
due to a data breach or temporarily made unavailable due to a natural disaster are
real concerns.
e Security challenges for cloud service customers :
1. Ambiguity in responsibility : A CSC uses services based on different service
categories as well as different deployment models. If the responsibilities are
not clearly defined in any of these cases then it may result in inconsistency or
may leave an open gate for attacks.
2. Loss of trust : Because of the abstraction of the security implementation details
between a CSC and a CSP, it is difficult for a CSC to get details of the security
mechanisms that the CSP has implemented to keep the cloud data secure.
3. Loss of governance : When the CSC uses cloud services, it has to move its
data onto the cloud and has to provide certain privileges to the CSP for
handling the data in the cloud. This may result in misconfiguration or an
attack due to the abstraction of the CSP's cloud practices and due to the
privileges that need to be given to the CSP.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 5-7 Security in Cloud Computing

4. Loss of privacy : CSC's privacy may be violated due to leakage of private


information while the CSP is processing CSC's private data or using the private
information for a purpose that the CSP and CSC haven't agreed upon.
5. Cloud service provider lock-in : This issue arises if a CSP doesn't abide by the
standard functions or frameworks of cloud computing and hence makes it
difficult for a CSC using its services to migrate to any other CSP. The use of
non-standard functions and cloud framework makes the CSP
non-inter-operable with other CSPs and also leaves CSC open to security
attacks.
6. Misappropriation of intellectual property : A CSC may face this challenge due
to the possibility that a CSC's data on the cloud might leak to third parties that
are using the same CSP for their cloud services. This leakage may violate the
CSC's copyrights and may result in the disclosure of CSC's private data.
7. Loss of software integrity : A CSC encounters this challenge due to the fact
that its software is running in the cloud once it is given to the CSP. It is
possible that this software might be tampered with or might be affected while
the software is running in the CSP and is not in CSC's control, resulting in
CSC's loss over its software.

EX] Advantages
Data centralization : service provider takes responsibility of storage and small
organization need not spend more money for personal storage device.
Incident response : IaaS providers contribute dedicated legal server which can be
used on demand.
Forensic image verification time.
Logging : storage requirement for benchmark logs is mechanically solved.

[EXE] Disadvantages
Loss of control : The enterprise's loss of control in enhancing the network's
security is the most significant disadvantage of cloud computing security. The
responsibility of securing the network is shared between the Cloud Service
Provider (CSP) and the enterprise.
Reduced visibility and control : when migrating to a cloud based computing
model, organizations will lose a degree of visibility and control, with some
responsibility for policies and infrastructure moving to the cloud provider.
Unsecure API and interfaces.

Data segregation

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 5-8 Security in Cloud Computing

1. What are the security challenges in cloud computing ?


SPPU : March-20, In Sem, Marks 5

5.5 | Content Level Security


¢ Content-based security is a departure from traditional enterprise content
management security measures that focus on restricting access to a static
repository or network, or on securing specific devices or applications.
¢ Specific content-based security features include restricting who can open, email,
print or edit a piece of content and placing a time limit on how long a user can
access a given piece of content. Content can expire from a given repository and no
longer be viewable by anyone.
¢ Content-based security enables custodians of enterprise information to define and
control the scope of actions available for users handling content (such as business
records or documents), regardless of the physical location of the content in
question. This can be useful for organizations that extensively use cloud
computing and enterprise mobility technologies that take company information
outside the enterprise firewall.

EX] Cloud Security Services


e The basic security services for information security include assurance of data
confidentiality, integrity and availability.

¢ Fig. 5.6.1 shows organization of data security and privacy in cloud computing.

dN
I Aly
T=
/ >| Hardware
td

Fig. 5.6.1

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 5-9 Security in Cloud Computing

1. Confidentiality :
Confidentiality refers to limiting information access. Sensitive information should
be kept secret from individuals who are not authorized to see the information. In
cloud environments, confidentiality primarily pertains to restricting access to data
in transit and storage.
Data confidentiality is important for users to store their private or confidential
data in the cloud. Authentication and access control strategies are used to ensure
data confidentiality.
The data confidentiality, authentication and access control issues in cloud
computing could be addressed by increasing the cloud reliability and
trustworthiness.
Because the users do not trust the cloud providers and cloud storage service
providers are virtually impossible to eliminate potential insider threat, it is very
dangerous for users to store their sensitive data in cloud storage directly.
Simple encryption is faced with the key management problem and cannot support
complex requirements such as query, parallel modification and fine-grained
authorization.
2. Integrity :
This service protects data from malicious modification. When having outsource
their data to remote cloud servers, cloud users must have a way to check whether
or not their data at rest or in transit are intact. Such a security service would be of
the core value to cloud users.
Integrity can extend to how data is stored, processed and retrieved by cloud
services and cloud-based IT resources.
Data integrity in the cloud system means preserving information integrity. The
data should not be lost or modified by unauthorized users.
Data integrity in the cloud system means preserving information integrity. The
data should not be lost or modified by unauthorized users.
Data integrity is the basis to provide cloud computing service such as SaaS, PaaS
and IaaS.
Besides data storage of large-scaled data, cloud computing environment usually
provides data processing service. Data integrity can be obtained by techniques
such as RAID-like strategies and digital signature.
3. Availability :
This service assures that data stored in the cloud are available on each user
retrieval request. This service is particularly important for data at rest in cloud
servers and related to the fulfillment of service level agreement.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 5-10 Security in Cloud Computing

Data availability means the following : When accidents such as hard disk damage,
IDC fire, and network failures occur, the extent that user's data can be used or
recovered and how the users verify their data by techniques rather than
depending on the credit guarantee by the cloud service provider alone.
The cloud service provider should ensure the data security, particularly data
confidentiality and integrity. The cloud provider should share all such concerns
with the client and build trust relationship in this connection. The cloud vendor
should provide guarantees of data safety and explain jurisdiction of local laws to
the clients.
Disaster recovery plan is a plan designed to recover all the vital business processes
during a disaster with in a limited amount of time. This plan has all the
procedures required to handle the emergency situations.
A disaster recovery process should have provable recovery capability, and hence it
provides the most efficient method to be adopted immediately after a disaster
occurs.

Security Authorization Challenges in the Cloud


Authorization is the function of specifying access rights/privileges to resources
related to information security and computer security in general and to access
control in particular.
Authorization determines what the user can access and what he cannot access
1. Auditing :
Cloud security audit can help by assessing and prioritizing risks, evaluating
current controls, identifying the gaps in existing cloud security strategy and
programs and making recommendations tied to business priorities.
Functions performed by IT auditors :
a. Backup controls
b. Data center security
c. System development standards
d. System and transaction controls
e. Contingency plan.

2. Accountability :
This is the process that keeps track of a user's activity while attached to a system;
the trail included the amount of time attached, the resources accessed, and how
much data transferred.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 5-11 Security in Cloud Computing

Accounting data is used for trending, detecting breaches and forensic


investigating. Keeping track of users and their activities serves many purposes.
For example, tracing back to events leading up to a cyber security incident can
prove very valuable to a forensics analysis and investigation case.

EX] secure Cloud Software Requirements


Requirements of secure cloud software are as follows :
1. Secure Development Practices : It includes data handling, code practices,
language options, input validation and content injection, physical security of
the system.
2. Approaches to cloud software requirements engineering : A resource
perspective on cloud software security requirements, goal-oriented software
security requirements and monitoring internal and external requirements.
3. Cloud security policy implementation and decomposition : Includes
implementation issues, decomposing critical security issues into secure cloud
software requirements (Confidentiality, Integrity, Availability, Authentication
and Identification, Authorization, Auditing).

5.9 | Secure Cloud Software Testing


Cloud testing, also called cloud-based testing, is the assessment of a Web
application's performance, reliability, scalability and security in a third-party's
cloud computing environment.
Compared to a traditional on-premises environment, cloud testing offers users
pay-per-use pricing, flexibility and reduced time-to-market.
The test processes and technologies used to perform functional testing against
cloud-based applications are not significantly different than traditional in-house
applications, but awareness of the non-functional risks around the cloud is critical
to success.
For example, if testing involves production data, then appropriate security and
data integrity processes and procedures need to be in place and validated before
functional testing can begin.
In the cloud environment, any application can be subjected to the following types
of testing :
1. Functional testing to ensure that software meets functional requirements
2. Non-functional testing to ensure the quality of service
3. Ability testing to show whether users will receive application services from the
cloud environment on-demand

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 5-12 Security in Cloud Computing

¢ Functional testing : Functional software testing checks all the features and
functions of software and its interaction with hardware. For conducting functional
testing, testers can use such tools as Rapise, Sauce Labs and TimeShiftX

e Non-functional testing : Non-functional testing is also known as performance


testing, as it allows you to check the non-functional aspects of software like its
performance, usability, and reliability. For conducting this type of testing, you can
use cloud-based tools such as CloudTest, AppPerfect, CloudTestGo and
AppLoader
o Ability testing : Ability testing is necessary to verify whether users really receive
application services on demand.
¢ Cloud testing focuses on the core components like
1. Application : It covers testing of functions, end-to-end business workflows, data
security, browser compatibility, etc.
2. Network : It includes testing various network bandwidths, protocols and
successful transfer of data through networks.
3. Infrastructure : It covers disaster recovery test, backups, secure connection and
storage policies. The infrastructure needs to be validated for regulatory
compliances.

Types of testing in
the cloud
|
I I ]
Functional Non-functional i”
testing testing Ability testing

; Business Compatibility and


1 System testing —| requirement 1 interoperability
testing testing

Acceptance . ) Disaster
g 1 Security testing 1 recovery
ising testing

|| Integration L | sell ly Ene Multi-tenancy


testing periormance | testing
testing

Fig. 5.9.1

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 5-13 Security in Cloud Computing

5.9.1 | Type of Testing in Cloud

The whole cloud testing is segmented into four main categories,


a) Testing of the whole cloud : The cloud is viewed as a whole entity and based
on its features testing is carried out. Cloud and SaaS vendors, as well as end
users, are interested in carrying out this type of testing
b) Testing within a cloud : By checking each of its internal features, testing is
carried out. Only cloud vendors can perform this type of testing
c) Testing across cloud : Testing is carried out on different types of cloud-like
private, public and hybrid clouds
d) SaaS testing in cloud : Functional and non-functional testing is carried out on
the basis of application requirements

5.9.2 | Benefit of Cloud-based Testing

In contrast to traditional software testing, cloud-based testing has several unique


advantages :
Scalability : Cloud computing allows testers to increase or decrease computing
resources according to their needs.
Cost-cutting : In cloud computing, you pay only for those resources that you use.
Timesaving : With cloud-based testing, an application can be simultaneously run
on different hardware so testers can spend more time fixing defects.
Easily customizable : By using cloud-based tools and services, testers can easily
emulate an end-user-centric environment with minimum cost and time.
Properly configured test environment : It usually takes much time to properly set
up a test environment on multiple devices.
Ensure comprehensive testing : In order to conduct comprehensive testing, the test
team needs to run an application on all possible devices that support different
platforms, operating systems and browsers.
Faster testing : Cloud-based testing tools ensure automated testing, which greatly
reduces the time to market for software.
Constant availability : Software testing in the cloud is available to testers at any
time.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 5-14 Security in Cloud Computing

[EXT] Multiple Choice Questions


Ql ensures that information is not changed or altered in transit.

[al Integrity [b] Authentication


Confidentiality [d| Availability
Q.2 prevents either sender or receiver from denying a transmitted message.

[a] Integrity [b] Nonrepudiation


Confidentiality [d] Availability
Q3 management is the total process used to identify, control, and minimize
the impact of uncertain events.
[a] Software [b] Hardware
Risk [d] All of these

Q.4 The responsibility of securing the network is shared between the _______ and the
enterprise.

[a] network user [b] cloud service provider


middle service provider [d] all of these

Q.5 A security _____ is a statement produced by the senior management of an


organization.

[a] mechanism [b] policy


method [d| all of these

Answer Keys for Multiple Choice Questions :

Q1| a Q2| b Q3| ¢ (Q4| b [Q5| b

aaa

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Unit - VI

Advanced Techniques in
Cloud Computing

Syllabus
Future Trends in cloud Computing, Mobile Cloud, Automatic Cloud Computing : Comet Cloud.
Multimedia Cloud : IPTV, Energy Aware Cloud Computing, Jungle Computing, Distributed Cloud
Computing Vs Edge Computing, Containers, Docker, and Kubernetes, Introduction to DevOps.
IOT and Cloud Convergence : The Cloud and IoT in your Home, The IOT and cloud in your
Automobile, PERSONAL : IoT in Healthcare.

Contents
6.1 Future Trends in Cloud Computing
6.2 Mobile Cloud
6.3 Automatic Cloud Computing. .............. June-19,--------r-ii Marks 8
6.4 Multimedia Cloud
6.5 Energy Aware Cloud Computing ........... June-19,-- «csi Marks 8
6.6 Jungle Computing
6.7 Docker LL. June-19, Dec.-19, -------- Marks 8
6.8 Introduction to DevOps
6.9 IOT and Cloud Convergence
6.10 Multiple Choice Questions

(6-1)
Cloud Computing 6-2 Advanced Techniques in Cloud Computing

xi Future Trends in Cloud Computing


The future of Cloud computing as a combination of cloud-based software products
and on-premises compute which will help to create hybrid IT solutions.
Multi-cloud refers to the presence of more than one cloud service (public or
private). The service is sourced from different cloud vendors.
Multi-cloud distributes workloads across multiple cloud environments. You can
use a private cloud for storing sensitive data. At the same time, host less sensitive
workloads on the public cloud.
When a combination of private as well as a third-party public cloud service is
used, this is known as Hybrid cloud computing. The primary use of this type of
cloud computing is to allow you to shift your workload between the private and
public clouds, thus, allowing you to have more flexibility. This is one example of
how a hybrid/multi-cloud solution can be used to reduce a users workload.
With the help of the Internet of Things, the quality of the internet can be
increased. With the help of the IoT and Cloud Computing, we can store data in
the cloud, for further analyze and provide enhanced performance.

[XEN Just Enough Operating System


Just Enough Operating System (JeOS) is a tech design concept in which a leaner
version of an OS replaces the full version to run on a specific device or hardware
setup.
Users of cloud based software access applications via a browser. They thus do not
need an OS to do more than run the browser.
Cloud systems share memory, disk space, peripherals and other resources among
multiple applications and users by providing each with isolated access to the
network, storage, and server virtualization layers.

The systems use a hypervisor to isolate allocated hardware resources to run


workloads.
In contrast to the traditional operating system, a JeOS offers only the components
necessary to make an application run, without unnecessary interfaces, functions,
libraries, and services.
It is packaged with the software to form a prebuilt, preconfigured, ready to-run
application appliance within a virtual machine.
The JeOS is a lightweight, low memory-footprint, easy-to-install system including
only the code necessary to boot and support a file system, storage devices, and
networking.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 6-3 Advanced Techniques in Cloud Computing

Ubuntu Linux-based OS, which has a JeOS version. Other JeOS platforms include
Novell's SUSE Linux Enterprise JeOS, Oracle’s Enterprise Linux JeOS, and Sun's
OpenSolaris JeOS.

IX] Mobile Cloud


One of the main benefits of cloud computing is reducing downtime and wasted
expenditure for servers and other computer equipment. A given company is
required to purchase the minimum amount of hardware necessary to handle the
maximum points of stress on their system.
Given situations where the strain and traffic are highly variable this leads to
wasted money. For example, Amazon.com, a pioneer in cloud computing, at times
used as little as 10% of their capacity so that they would have enough capacity to
deal with those rarer high strain times.
Mobile Cloud Computing (MCC) at its simplest, refers to an infrastructure where
both the data storage and data processing happen outside of the mobile device.
Fig. 6.2.1 shows block diagram
of mobile cloud.

me
Mobile cloud applications —_— Ry —
move the computing power A @ -—
and data storage away from

ne
mobile phones and into the
Smart mobile Wireless network Computational
cloud, bringing applications device technology cloud
and mobile computing to not
just smart phone users but a Fig. 6.2.1 Block diagram of mobile cloud
much broader range of mobile
subscribers.
Mobile cloud applications move the computing power and data storage away from
the mobile devices and into powerful and centralized computing platforms located
in clouds, which are then accessed over the wireless connection based on a thin
native client.
Mobile devices face many resource challenges (battery life, storage, bandwidth
etc.).
Cloud computing offers advantages to users by allowing them to use
infrastructure, platforms and software by cloud providers at low cost and
elastically in an on-demand fashion.
Mobile cloud computing provides mobile users with data storage and processing
services in clouds, obviating the need to have a powerful device configuration (e.g.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 6-4 Advanced Techniques in Cloud Computing

CPU speed, memory capacity), as all resource-intensive computing can be


performed in the cloud.
o Fig. 6.2.2 shows mobile cloud computing architecture.

Mobile networkA

Bog Mobile network services


Data center owners or

R Hh
|}
cloud service providers

ol

Cloud computing

Mobile network B

Mobile network services

Hh
|}
ME
i point
| Contral Servers

et A
Application
processors
Internet service servers
providers (ISPs)
Mobile | BTS
devices 1

T T
1 Application service providers

Mobile users Network operators

Fig. 6.2.2 Mobile cloud computing architecture

In mobile cloud computing mobile network and cloud computing are combined,
thereby providing an optimal services for mobile clients.
Cloud computing exists when tasks and data are kept on individual devices.
Applications run on a remote server and then sent to the client.
Here the mobile devices are connected to the mobile networks through the base
stations; they will establish and control the connections (air interface) and
functional interfaces between the mobile networks and mobile devices.
Mobile users send service requests to the cloud through a web browser or desktop
application. The informations are transmitted to the central processors that are
connected to the servers providing mobile network services.
Here, services like AAA (Authentication, Authorization and Accounting) can be
provided to the users based on Home Agent (HA) and subscriber's data stored in
databases

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 6-5 Advanced Techniques in Cloud Computing

Mobile devices are connected to the mobile networks via base stations that
establish and control the connections and functional interfaces between the
networks and mobile devices.

Mobile users’ requests and information are transmitted to the central processors
that are connected to servers providing mobile network services.
The subscribers’ requests are delivered to a cloud through the Internet.
In the cloud, cloud controllers process the requests to provide mobile users with
the corresponding cloud services.

6.2.1 | Advantages and Disadvantages


Advantages :
1. Saves battery power
2. Makes execution faster
3. Improves data storage capacity and processing power
4. Improves reliability and availability : Keeping data and application in the clouds
reduces the chance of lost on the mobile devices
Dynamic provisioning : Dynamic on-demand provisioning of resources on a
fine-grained, self-service basis.

Disadvantages :
1. Must send the program states (data) to the cloud server
2. Network latency can lead to execution delay.

XX] Mobile Cloud Application


Mobile Gaming ® M-game is a high potential market generating revenues for service
providers.
® Can completely offload game engine requiring large computing
resource (e.g., graphic rendering) to the server in the cloud.
® Offloading can also save energy and increase game playing time (eg.
MAUI allows fine-grained energy-aware offloading of mobile codes to
a cloud)

® Rendering adaptation technique can dynamically adjust the game


rendering parameters based on communication constraints and gamers’
demands
Mobile Healthcare ® M-healthcare is to minimize the limitations of traditional medical
treatment (eg. Small storage, security/privacy, medical errors, ...)

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 6-6 Advanced Techniques in Cloud Computing

® M-healthcare provides mobile users with convenient access to


resources (eg. medical records)
® M-healthcare offers hospitals and healthcare organizations a variety of
on-demand services on clouds
Mobile Learning ® M-learning combines e-learning and mobility
® Traditional m-learning has limitations on high cost of
devices/network, low transmission rate, limited educational resources

® Cloud-based m-learning can solve these limitations

® Enhanced communication quality between students and teachers


® Help learners access remote learning resources
Mobile Commerce ® M-commerce allows business models for commerce using mobile
devices.
® Examples : Mobile financial, mobile advertising, mobile shopping.

® M-commerce applications face various challenges


® Integrated with cloud can help address these issues

® Example : Combining 3G and cloud to increase data processing speed


and security level.

[EX] Automatic Cloud Computing


Autonomic computing is the ability of distributed system to manage its resources
with little or no human intervention. It involves intelligently adapting to
environment and requests by users in such a way the user does not even know.
Autonomic monitoring are mostly implemented on specific layers of the cloud
computing architecture.
Fig. 6.3.1 shows the high-level architecture enabling autonomic management of
SaaS applications on Clouds.
SaaS application portal : This component hosts the SaaS application using a Web
Service-enabled portal system.
Users or brokers acting on their behalf submit service requests from anywhere in
the world to these SaaS applications.
Autonomic management system and PaaS framework : This layer serves as a
platform as a service. Its architecture comprises of autonomic management
components to be integrated in the PaaS level, along with modules enforcing
security and energy efficiency.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 6-7 Advanced Techniques in Cloud Computing

- Ny g ~ 73
» Ax = | :
3 < Saa$ applications portal 7 my

Healthcare Spatial-temporal analytics

SaaS/Paasl integration layer


T
ee ad —-===~
f Autonomic management system A \
1 - - 1
1 ( Security and attack detection | Workflow 1
a
|| oo -User QoS K———
XRD0-0 J
|
S 1 Application ~~ -Overall cost |
1 scheduler ~~ -Energy efficiency 1
" -Data location > PaaS framework |

: Dynamic resource provisioning algorithms Integration ;


Pe ———————————— mm —

PaaS/laa8S integration layer

Private cloud Public clouds

EX

Fig. 6.3.1 System architecture for autonomic cloud management

e User QoS-based application scheduler and dynamic resource provisioning


algorithms are added as plug-ins.
¢ Infrastructure as a service : This layer comprises distributed resources provided by
private and public clouds.
e SaaS is described as a software application deployed as a hosted service and
accessed over the Internet.
e In order to manage the SaaS applications in large scale, the PaaS layer has to
coordinate the cloud resources according to the SaaS requirements, which is
ultimately the user QoS.
¢ The autonomic management system incorporates the following services in the PaaS
layer : Security and attack detection, application scheduling and dynamic
provisioning.
¢ The autonomic manager is composed by the following components, with specific
roles :

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 6-8 Advanced Techniques in Cloud Computing

Application scheduler : The scheduler is responsible for assigning each task in an


application to resources for execution based on user QoS parameters and the
overall cost for the service provider.
Energy-efficient scheduler : One of the main objectives to be optimized during
the application scheduling process is energy utilization. Applications need to be
scheduled in resources in such a way that their total energy consumption is
minimized. However, the algorithm has to achieve this goal without compromising
SLAs and cost.
Dynamic resource provisioning algorithms : This component implements the logic
for provisioning and managing virtualized resources in private and public Cloud
environments based on the resource requirements as directed by the application
scheduler.
Security and attack detection : This component implements all the checks to be
performed when requests are received in order to evaluate their legitimacy. This
prevents the scaling-up of resources to respond to requests created with the
intention of causing a Denial of Service or other forms of cyber-attacks.
The module must be able to distinguish between authorized access and attacks,
and in case of suspicion of attack, it can either decide to drop the request or avoid
excessive provision of resources to it.

6.3.1 | CometCloud

CometCloud is based on a decentralized coordination substrate, and supports


highly heterogeneous and dynamic cloud/Grid infrastructures, integration of
public/private clouds and cloudbursts.
CometCloud is an autonomic computing engine for cloud and grid environments.
CometCloud is composed of a programming layer, a service layer, and an
infrastructure layer.
Fig. 6.3.2 shows CometCloud architecture for autonomic cloudbursts.
The infrastructure layer uses the Chord self-organizing overlay, and the Squid
information discovery and content-based routing substrate built on top of Chord.
The routing engine supports flexible content-based routing and complex querying
using partial keywords, wildcards, or ranges.
This layer also provides replication and load balancing services, and it handles
dynamic joins and leaves of nodes as well as node failures.
The service layer provides a range of services to supports autonomics at the
programming and application level. An application can switch between spaces at
runtime and can simultaneously use multiple spaces.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 6-9 Advanced Techniques in Cloud Computing

Application

Master/Worker/BOT
Programming MaReduce/
layer ) oo Task Workflow Hadoop
Scheduling Monitoring :
consistency

Clustering/ —
Service Anotomy detection Coordination Publish/Subscribe
layer
Discovery Event Messaging

Replication Load balancing

acre Content-based routing Content security

Self-organizing layer

Data center/Grid/Cloud

Fig. 6.3.2 CometCloud architecture for autonomic cloudbursts

¢ This layer also provides asynchronous (publish/subscribe) messaging and evening


services.
e The programming layer provides the basic framework for application development
and management. It supports a range of paradigms including the
master/worker/BOT. Masters generate tasks and workers consume them.

1. Describe cloudlets for mobile cloud computing with neat diagram and differentiate between
cloudlets and clouds. SPPU : June-19, Fnd Sem, Marks 8

XA Multimedia Cloud
e Due to the invention of cloud computing, nowadays users can easily access the
multimedia content over the internet at any time. User can efficiently store the
multimedia content of any type and of any size in the cloud after subscribing it
with no difficulties.
¢ Not only storing the media content like Audio, Video and Image, but can process
them within the cloud since the computation time for processing media data is
more in complex hardware.
e After processing the processed data can be easily received from the cloud through
a client without any need of installing complex hardware.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 6-10 Advanced Techniques in Cloud Computing

e Fig. 6.4.1 shows fundamental concept of multimedia cloud.

Audio deo Image

CA = 0 0 0 Media cloud

2
Storage
E—

CPU

Fig. 6.4.1 Fundamental concept of multimedia cloud


¢ Thus multimedia cloud computing is the processing, accessing and storing of
multimedia contents like audio, video and image using the services and
applications available in the cloud without physically acquiring them.
¢ Currently many company’s clouds like AmazonEC2, Google Music, DropBox,
SkyDrive provides content management system within the cloud network.
e The users of these clouds can access the multimedia content for example; the user
can view a video anywhere in the world at anytime using their computers, tablets
or smart phones.
e Cloud media is a cloud which has the multimedia content of the owner of that
particular cloud. The media content can be accessed through the multimedia
signaling protocols in the cloud and can be streamed to clients present in
computers, tablets, cars and smart phones.
¢ Fig. 6.4.2 shows relation between cloud media and media cloud.

Cloud media
Authoring/editing Sharing/streaming
service service

MSPs

mmm ———a Media cloud

Hard disk — Resource allocator

1
oP0 K——> Load balancer

Fig. 6.4.2 Relation between cloud media and media cloud

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 6-11 Advanced Techniques in Cloud Computing

Not only processing, but the media content can be shared between clouds using
the streaming protocols like TCP/IP, UDP, RTP, HTTP etc.
Streaming of media content involves, loading or buffering media data, coding,
mixing, rating and rendering over the service providers.
Other profiling, packetizing, tokenizing of media contents will be done by the
cloud based on the streaming protocols used and it will be streamed to the client
system.
Cloud media technology offers number of key benefits to its service providers as
well as the users through increased implementation time, efficient data storage
capacity, less computation and cost.
It created a striking impact in the multimedia content processing like editing,
storing, encrypting and decrypting, gaming, streaming, compressing etc.

6.4.1 Hla"
Today, consumers watch video on a variety of connected devices. New
Over-The-Top (OTT) providers such as Netflix are offering direct-to-consumer
services with low prices, advanced user interfaces and easy access to multi-screen
video.
Changing usage patterns brought on by subscriber desire to watch content at the
time, location and on the device of their choosing are increasing content
distribution costs.
Pay TV providers are particularly susceptible to these trends and need to adapt
their traditional TV delivery architectures to offer innovative services that attract
and retain customers.
The traditional Set-Top Box (STB) will disappear. The functions of today’s STB
hardware will be carried out in the network and by the connected device itself,
eliminating the cost and complexity of managing home-based STBs.
Traffic will be all unicast. Over time, device format fragmentation, time-shifting
viewing habits and service personalization will erode broadcast and multicast
efficiencies.
Ultimately, every end user will be served with a unique stream. Services will be
deployed in the cloud.
Dedicated video platforms will migrate to cloud-based services, reducing costs and
accelerating time to market.
Operators will move from vertically integrated middleware stacks to more open
architectures with best-of-breed components.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 6-12 Advanced Techniques in Cloud Computing

Cloud DVR technology makes all TV content available on demand, on any device
and in any location.
There are several advantages of cloud-based services-the key benefits are that the
service is software-based, so one doesn’t need a physical location to run the
operations. As a result, real estate, infrastructure and manpower cost reduce
dramatically.
For instance, if a Bollywood film channel wants to check if it has a market in the
US, it can use the cloud-based technology to launch without adding to its cost by
leasing a satellite for a minimum of three to five years.
A product of ActiveVideo, a Silicon Valley software company, CloudTV is
available on more than 15 million devices.

XA Energy Aware Cloud Computing


The energy consumption of cloud computing continues to be an area of significant
concern as data center growth continues to increase.

6.5.1 | Green Cloud

Cloud computing is a highly scalable and cost-effective infrastructure for running


HPC, enterprise and Web applications.
However, the growing demand of cloud infrastructure has drastically increased the
energy consumption of data centers, which has become a critical issue.
energy-efficient solutions are required to minimize the impact of cloud computing
on the environment.
Data centres are not only expensive to maintain, but also unfriendly to the
environment.
Cloud service providers need to adopt measures to ensure that their profit margin
is not dramatically reduced due to high energy costs.
Amazon.com's estimate the energy-related costs of its data centers amount to 42 %
of the total budget that include both direct power consumption and the cooling
infrastructure amortized over a 15-year period.
Google, Microsoft, and Yahoo are building large data centers in barren desert land
surrounding the Columbia River, USA to exploit cheap hydroelectric power.
There is also increasing pressure from Governments worldwide to reduce carbon
footprints, which have a significant impact on climate change.
As energy costs are increasing while availability dwindles, there is a need to
shift focus from optimising data centre resource management for pure

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 6-13 Advanced Techniques in Cloud Computing

performance alone to optimising for energy efficiency while maintaining high


service level performance.
Green cloud computing model that achieves not only efficient processing and
utilisation of computing infrastructure, but also minimise energy consumption.
Exponential data growth leads to greater storage needs. The traditional approach
of adding more disks and storage systems satisfies demand, but also increases
power, cooling, and space requirements.
Green storage is the practice of using a variety of “clean energy” storage methods
and products to cut down on a data center's carbon footprint, as well as cost.
There are a number of systems that can be used for green storage, and by
analyzing your data center and specific needs, you can choose a combination of
solutions that work for your environment :
. Tape is a popular method of green storage that is widely used. Tape has no moving
parts that use up energy, is portable and has a longer shelf-life than other storage
technology.
. Virtualized servers can host up to 20 virtualized servers on one physical server.
This improves efficiency and cuts down on the need for expensive hardware.
. Not as widely used, but growing in popularity, are solid-state drives (SSDs),
which are energy efficient and faster than mechanical hard disk drives. However,
SSDs come at a higher cost than other methods.
. A massive array of idle disks (MAID) system only spins active drives, cutting
down on energy use and prolonging shelf-life. This architecture has been around
for a while but hasn't widely caught on largely.

[XE] Green Computing


Computers today are an integral part of individuals’ lives all around the world;
but unfortunately these devices are toxic to the environment given the materials
used, their limited battery life and technological obsolescence.
Green IT refers to the study and practice of designing, manufacturing, using, and
disposing of computers, servers, and associated subsystems such as monitors,
printers, storage devices, and networking and communications systems efficiently
and effectively with minimal or no impact on the environment.
Green computing refers to the practice and procedures of using computing
resources in an environment friendly way while maintaining overall computing
performance.
Green computing is the environmentally responsible and eco-friendly use of
computers and their resources.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 6-14 Advanced Techniques in Cloud Computing

¢ Computers and other IT infrastructure consume significant amounts of electricity,


which is increasing day by day, placing a heavy burden on our eleciric grids and
contributing to greenhouse gas (GHG) emissions.
e Green IT, also known as green computing.
¢ To promote green computing concepts at all possible levels, the following four
complementary approaches are employed :
1 Green use : Minimizing the electricity consumption of computers and their
peripheral devices and using them in an eco-friendly manner
2 Green disposal : Re-purposing an existing computer or appropriately disposing
of, or recycling, unwanted electronic equipment
3 Green design : Designing energy-efficient computers, servers, printers,
projectors and other digital devices
4 Green manufacturing : Minimizing waste during the manufacturing of
computers and other subsystems to reduce the environmental impact of these
activities.
e “Electronic waste” may be defined as discarded computers, office electronic
equipment, entertainment device electronics, mobile phones, television sets, and
refrigerators. This includes used electronics which are destined for reuse, resale,
salvage, recycling, or disposal.
e Green computing represents a responsible way to address the issue of global
warming. By adopting green computing practices, business leaders can contribute
positively to environmental stewardship and protect the environment while also
reducing energy and paper costs.
Benefit of Green IT
e Green IT benefits the environment by improving energy efficiency, lowering GHG
emissions, using less harmful materials and encouraging reuse and recycling.
1. Reduced in power and resource consumption
2. Green technology helps manage and recycle waste material
3. Reduced environmental impact and carbon footprint
4 . Improved operational efficiency.

6.5.3 | Energy-Saving Software Techniques

¢ Reducing power consumption is a challenge to system designers. Portable systems,


such as laptop computers and Personal Digital Assistants (PDAs) draw power
from batteries; so reducing power consumption extends their operating times.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 6-15 Advanced Techniques in Cloud Computing

e For desktop computers or servers, high power consumption raises temperature


and deteriorates performance and reliability. Power consumed by the CPU is
significant.
eo The total energy consumed by the system per cycle is the sum of
energies consumed by the processor and L1 cache, interconnects and pins,
memory, L2 cache, the DC-DC converter and the efficiency losses in the
battery.
¢ Energy-saving software techniques are as follows :

Techniques Sub-types

Computational Efficiency 1. Algorithm


2. Multithreading
3. uArch tuning
4. Vectorination

Data Efficiency 1. Caching


2. Asynchronous I/O
Context Awareness 1. AC/DC

2. Policy for power


3. Thresholds
Idle Efficiency 1. Timer resolution
2 . Background activity
3. C-states

6.5.4 | Key Issue Related to Cloud Computing Energy Efficiency

¢ Objective of computing by business providers for cloud.


e Cost-wise advantage of public cloud computing provides over traditional data
centers.
e Improvement of sustainability while shifting to the cloud.
e Impact of using cloud computing on carbon emission from the data center
operations.
¢ By migrating to the cloud, industries can achieve significant energy saving and
reduced pollution.
e The reduction in energy consumption was larger and not by a reduced number of
servers.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 6-16 Advanced Techniques in Cloud Computing

1. What is energy aware cloud computing ? Explain in detail.


SPPU : June-19, End Sem, Marks 8

6.6 | Jungle Computing


¢ Jungle computing is distributed computing system.
e A Jungle computing system consists of all compute resources available to
end-users, which includes clusters, clouds, grids, desktop grids, supercomputers,
as well as stand-alone machines and even mobile devices.
¢ Reasons for using Jungle Computing Systems :
1. An application may require more compute power than available in any one
system a user has access to.
2. Different parts of an application may have different computational
requirements, with no single system that meets all requirements.
e From a high-level view, all resources in a Jungle Computing System are in some
way equal, all consisting of some amount of processing power, memory and
possibly storage.
e End-users perceive these resources as just that : A compute resource to run their
application.
e¢ When grid computing was introduced over a decade ago, its foremost visionary
aim was to provide efficient and transparent socket computing over a distributed
set of resources.
e Many other distributed computing paradigms have been introduced, including
peer-to-peer computing, volunteer computing and more recently cloud computing.
e These paradigms all share many of the goals of grid computing, eventually aiming
to provide end-users with access to distributed resources with as little effort as
possible.
¢ These new distributed computing paradigms have led to a diverse collection of
resources available to research scientists, which include stand-alone machines,
cluster systems, grids, clouds, desktop grids, etc.

e With clusters, grids and clouds thus being equipped with multi-core processors
and many-core ‘add-ons’, systems available to scientists are becoming increasingly
hard to program and use.
e Despite the fact that the programming and efficient use of many-cores is known to
be hard, this is not the only problem. With the increasing heterogeneity of the
underlying hardware, the efficient mapping of computational problems onto the

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 6-17 Advanced Techniques in Cloud Computing

‘bare metal’ has become vastly more complex. Now more than ever, programmers
must be aware of the potential for parallelism at all levels of granularity.

Docker SPPU : June-19, Dec. 19


e Docker is quickly changing the way that organizations are deploying software at
scale.
e Docker is a tool that promises to easily encapsulate the process of creating a
distributable artifact for any application, deploying it at scale into any
environment, and streamlining the workflow and responsiveness of agile software
organizations.
¢ Benefits:
1. Packaging software in a way that leverages the skills developers already have.
2. Bundling application software and required OS file systems together in a single
standardized image format
3. Abstracting software applications from the hardware without sacrificing
resources

Process Simplification
e Docker can simplify both workflows and communication, and that usually starts
with the deployment story.
e Fig. 6.7.1 shows workflow with and without docker.

1
Development team Operations team
i
Request resources : Resources provisioned
I
' i
Script the deployment Tweak the deployment
T
v i1
T

Tweak the deployment Loop N times

Discover a dependency Install dependency

Application is deployed

Fig. 6.7.1 Traditional deployment workflow (without Docker)

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 6-18 Advanced Techniques in Cloud Computing

= . Application developers request resources from operations engineers.


. Resources are provisioned and handed over to developers.
WN

. Developers script and tool their deployment.


. Operations engineers and developers tweak the deployment repeatedly.
ok

. Additional application dependencies are discovered by developers.


Ul

. Operations engineers work to install the additional requirements.


NON

. Go to step 5 and 6
. The application is deployed.
00

Fig. 6.7.2 shows Docker deployment workflow.


®

Development team Operations team

Build image

Ship image to registry Provide configuration information

1
Deploy

Application is deployed

Fig. 6.7.2 Docker deployment workflow

1. Developers build the Docker image and ship it to the registry.


2. Operations engineers provide configuration details to the container and provision
resources.
3. Developers trigger deployment.

Broad Support and Adoption

e Docker is increasingly well supported, with the majority of the large public clouds.
For example, Docker runs on AWS Elastic Beanstalk, Google AppEngine, IBM
Cloud, Microsoft Azure, etc.
e Google's Eric Brewer announced that Google would be supporting Docker as its
primary internal container format. Rather than just being good PR for these
companies, what this means for the Docker community is that there is starting to
be a lot of money backing the stability and success of the Docker platform.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 6-19 Advanced Techniques in Cloud Computing

When docker released their libswarm development library at docker-Con 2014, an


engineer from Orchard demonsirated deploying a docker container to a
heterogeneous mix of cloud providers at the same time.
The Docker-client runs directly on most major operating systems, but because the
Docker server uses Linux containers, it does not run on non-Linux systems.

Docker has traditionally been developed on the Ubuntu Linux distribution, but
today most Linux distributions and other major operating systems are now
supported where possible.

Architecture
The fundamental architecture of Docker is a simple client - server model, with
only one executable that acts as both components, depending on how you invoke
the docker command.
Underneath those simple exteriors, Docker heavily leverages kernel mechanisms
such as IPTABLES, virtual bridging, cgroups, namespaces, and various filesystem
drivers.
Fig. 6.7.3 shows docker architecture.

IF [i
ol [gf [gf 180—
= ——
‘a
| | ‘a al "2
IE

1 1
apes 1

= I =| ol 1°] 19 I nede ]
Client Docker daemon 1 TS 1
[Festos 1 i MMT
Server Docker host Docker registry

Fig. 6.7.3 Docker architecture

It consists of two parts : The client and the server. Registry is one more
components which stores docker images and metadata about those images.
Docker Engine is a client-server based application with following components -
1. A server which is a continuously running service called a daemon process.
2. A REST API which interfaces the programs to use talk with the daemon and
give instruct it what to do.
3. A command line interface client.
Docker client is the primary service using which docker users communicate with
the docker. When we use commands "docker run" the client sends these
commands to dockerd, which execute them out.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 6-20 Advanced Techniques in Cloud Computing

Docker server
on Linux host or Virtual machine

Docker registry
on Linux host, VM or SaaS

Docker client
on local workstation

Fig. 6.7.4 Data flow

¢ The command used by docker depend on docker AP. In docker client can interact
more than one daemon process.
e The docker images are building the block of docker or docker image is a read-only
template with instructions to create a docker container. Docker images are the
most build part of docker life cycle
e The server does the ongoing work of running and managing your containers, and
you use the client to tell the server what to do.
¢ The docker daemon can run on any number of servers in the infrastructure, and a
single client can address any number of servers.
e Clients drive all of the communication, but docker servers can talk directly to
image registries when told to do so by the client.
¢ Clients are responsible for directing servers what to do and servers focus on
hosting containerized applications.
e Docker registry keeps docker images. We can run our private registry.
¢ When we run the docker pull and docker run commands, the required images are
pulled from our configured registry directory.
e Using docker push command, the image can be uploaded to our configured
registry directory.

Container and Kubernetes


¢ A container image is a ready-to-run software package that includes everything a
program needs to execute, including the code and any run-times it needs,
application and system libraries, and default values for any important settings.
o Container orchestration is concerned with the management of container lifecycles,
particularly in large, dynamic environments. Container orchestration is used by
software teams to control and automate a variety of tasks on container
management.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 6-21 Advanced Techniques in Cloud Computing

e Container orchestration works in any context where containers are employed. It


can assist you in deploying the same program across several environments
without having to rewrite it.
e Kubernetes is an open-source container management platform that unifies a cluster
of machines into a single pool of compute resources. With kubernetes, you
organize your applications in groups of containers, which it runs using the Docker
engine, taking care of keeping your application running as you request.
e Kubernetes is an open source container orchestration platform that automates
many of the manual processes involved in deploying, managing, and scaling
containerized applications.
¢ Kubernetes was originally developed and designed by engineers at Google.
e The primary responsibility of kubernetes is container orchestration. That means
making sure that all the containers that execute various workloads are scheduled
to run physical or virtual machines.
e The containers must be packed efficiently following the constraints of the
deployment environment and the cluster configuration. In addition, kubernetes
must keep an eye on all running containers and replace dead, unresponsive, or
otherwise unhealthy containers.
¢ Kubernetes uses docker to run images and manage containers.
e Kubernetes allows several containers to work in harmony, reducing operational
burden. Interestingly, this includes docker containers. Kubernetes can be integrated
with the docker engine, and uses "Kubelets” to coordinate the scheduling of docker
containers.
¢ The docker engine runs the container image, which is created by running docker
build. The higher-level concepts (load balancing, service discovery, and network
policies) are controlled by kubernetes. When combined, both docker and
kubernetes can develop a modern cloud architecture. However, it should be
remembered the two systems, at their core, are fundamentally different.
e Fig. 6.7.5 shows kubernetes architecture.
eo Kubelet : This function runs on nodes, reads container manifests, and assures
defined containers have started and are running.
e Node : These perform the assigned tasks, with the kubernetes master controlling
them.
eo Master : This controls the kubernetes nodes and is the source of all task
assignments.
¢ Pod : When one or more containers are deployed ito one node. Containers in a
pod will share a host name, an IP address, IPC, and other resources.

e Replication controller : Controls the number of "identical" copies in a pod that


should be running in different locations on the cluster.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 6-22 Advanced Techniques in Cloud Computing

TT eer TN
(Neer

Containers Containers
Control plane =
User interface POD 1 POD 2

oo) API server


( Docker engine |

Controller manager

Containers Containers
Kubectl
— POD 1 POD 2

( Docker engine |

Fig. 6.7.5 Kubernetes architecture


e Service : This will decouple the work definitions from the pods. Service requests
are automatically sent to the right pod, regardless of location.
e Kubectl : The primary configuration tool for kubernetes.
¢ Kubernetes objects : These are persistent entities within the Kubernetes system.
They are used to represent the state of the cluster

1. Explain the client server architecture of docker ? What are network ports and unix sockets ?
SPPU : June-19, End Sem, Marks 8

2. Explain docker with respect to process simplification. Broad support and Adoption,
architecture. SPPU : June-19, End Sem, Marks 8

3. Explain traditional as well as docker deployment workflow, Client server architecture of


docker ? SPPU : Dec.-19, End Sem, Marks 8

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 6-23 Advanced Techniques in Cloud Computing

xl Introduction to DevOps
DevOps is a collaboration between Development and IT Operations to make
software production and deployment in an automated and repeatable way.
DevOps helps to increase the organization's speed to deliver software applications
and services. The word 'DevOps' is a combination of two words, Development’
and 'Operations.’
Developers are willing to push out software faster and faster. Operations, on the
other hand, knows that rapid-fire changes without proper safeguards could
destabilize the system, which goes directly against their charter.
DevOps should be used for large distributed applications such as eCommerce sites
or applications hosted on a cloud platform.
DevOps bridges the gap between development and operations, creating significant
efficiencies across the development and deployment of software. DevOps includes
a strong emphasis on automation, helping reduce the overall number of errors.

xi IOT and Cloud Convergence


Cloud computing in IoT works as part of a collaboration and is used to store IoT
data. The cloud is a centralised server containing computer resources that can be
accessed whenever required. Cloud computing is an easy travel method for the
large data packages generated by the IoT through the Internet.
The Internet of Things (IoT) is the network of physical objects i.e. devices, vehicles,
buildings and other items embedded with electronics, software, sensors, and
network connectivity that enables these objects to collect and exchange data.
The Internet of Things is the intelligent connectivity of physical devices driving
massive gains in efficiency, business growth and quality of life.
The Internet of Things refers to the capability of everyday devices to connect to
other devices and people through the existing Internet infrastructure. Devices
connect and communicate in many ways. Examples of this are smart phones that
interact with other smart phones, vehicle-to-vehicle communication, connected
video cameras, and connected medical devices. They are able to communicate with
consumers, collect and transmit data to companies, and compile large amounts of
data for third parties

[XX] The Cloud and loT in your Home


Home automation is the automatic control of electronic devices in your home.
These devices are connected to the Internet, which allows them to be controlled
remotely.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 6-24 Advanced Techniques in Cloud Computing

¢ Interconnected devices enable to intelligently monitor and control smart homes in


a future Internet of Things.
e Energy saving applications, for example, control indoor climate and electricity
usage by employing context information to switch off appliances (e.g., lights,
computers), reduce room temperature, close windows, or stop warm water
circulation.
e Home automation works on three levels :
1. Monitoring : Monitoring means that users can check in on their devices
remotely through an app. For example, someone could view their live feed
from a smart security camera.
2. Control : Control means that the user can control these devices remotely, like
planning a security camera to see more of a living space.
3. Automation : Finally, automation means setting up devices to trigger one
another, like having a smart siren go off whenever an armed security camera
detects motion.
e Example : Smart Lighting

[XXK] Smart Lighting


e Smart control the lights with automation signal system to save energy. Smart,
connected lighting is the next - generation energy - efficient LED products with
additional sensors to sense things such as occupancy and temperature.
Fr HH) Smart home appliances »)
Smart lighting
Air conditioning =~
——
[m 2)

pa
Thermostat

vl J
a rr

Smart living
| ——.
Fig. 6.9.1 Smart home

e In automatic light control system, Light Dependent Resistor (LDR) sensor is used
to detect bright /medium /dim /dark conditions.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 6-25 Advanced Techniques in Cloud Computing

It is simple enough to envision the addition of sensors and communications to


create that initial concept of smarter, more adaptive lighting. If people are present,
turn the lights on; if not, turn them off. Or use your smart phone to connect to the
lighting system and tune it to the desired brightness level or to a particular color.
Smart lighting is considered the one of the main solutions for energy reduction by
means of controlling lighting level according to desired need with minimum
energy consumption.
Smart - lighting systems utilize motion and light sensors for performing the
control algorithms.
The system uses motion and light sensors for detecting the surrounding
environment. There are lamps controlled with the specific lighting level in order to
supply the adequate amount of lighting required without affecting the user
visibility.
Certainly the required lighting level is strongly dependent on the weather
conditions. In clear weather at night might require more luminance than cloudy
one, due to the reflection from the clouds.

While during mist and foggy weathers require the highest possible lighting level,
as the visibility reaches its lowest. On snowy weather it might require an
intermediate level between clear and foggy.
During night it requires high lighting levels, while at day it needs just fade level
to provide guidance or turn off if the weather is clear. The lighting concentration
in the yard is affected by the above conditions.

6.9.2 | The IOT and Cloud in your Automobile

Today, users of IoT devices can evaluate engine performance, control air
temperature, and measure physical health indicators with only a few clicks.
Conventional perceptions of the automotive industry are radically changing with
IoT development. Predictive maintenance, Wi-Fi capabilities powered by
3G/4G/5G functionality, Car2Car connectivity, and advanced fleet management
are only a few examples of how IoT-based solutions are shaping the new
automotive age.
The automobile industry is one of the fastest-growing markets for IoT-based
solutions. The number of installed connectivity units in vehicles is likely to
increase by 67 % between 2018 and 2020.
Predictive maintenance technology is based on the use of IoT connectivity tools
that collect data on the performance of different parts, transfer that data to the
cloud in real time, and evaluate the risks of potential malfunction of a car's

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 6-26 Advanced Techniques in Cloud Computing

hardware or software. After information is processed, a driver is notified and


advised of any necessary service or repair to avoid potential incidents.
Fig. 6.9.2 shows battery working.

I=T=T=1= « = » I=T=T=1=
—_— Low battery — —_— [4]
predicted 5
Battery Alert sent Issue
conditions avoided

Fig. 6.9.2 Battery working

Predictive maintenance can facilitate vehicle use by both private owners and
dealerships with large fleets of vehicles. It enables end-users to get the right
information in advance. With IoT connectivity tools, you can forget about
unplanned stops or breakdowns during the ride.

[XXX] Remote Vehicle Diagnostics


Remote Vehicle Diagnostics Solution monitors the health of the vehicle, determines
the root cause of the problem / failure and provides real time information of
vehicle parameters to assess its performance against benchmarks.
The solution monitors the health of the electric vehicle, commercial vehicle, utility
vehicle and provides insight to field support staff to determine the root cause of
the problem. It also enables the customers to access information about the vehicle.
Commercial / Utility vehicles being driven across the country extensively over
time for various purposes are in need of a diagnostic check which is automated
through the offering.
By monitoring all the aspects of the car is easier to detect any problem in advance
by sending all sensor readings to a certified center where technicians and
engineers will apply their expertise to find and predict imminent failures of key
systems integrated in the vehicle.
Modern commercial vehicles support on board diagnostic standard. Next
generation vehicles will have sophisticated on-board connectivity equipment,
providing wireless network access to the vehicle for infotainment and other
telematics services. Fig. 6.9.3 shows remote vehicle diagnostics.
In vehicle, sensors connect to the vehicle terminal which is responsible for
collecting, storing, processing and reporting information and responding to
commands from supervision platforms.
The vehicle terminal consists of the microprocessor, data storage, GPS module,
wireless communication transmission module, real time clock and data
communication interface.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 6-27 Advanced Techniques in Cloud Computing

Deployed vehicle Remote diagnostics datacenter


Diagnostic data

OO
Reformulated code
® Fault detection and isolation e Data analysis
© Fault tolerant controllers ® Code reformulation
® Remote code update ® Verification profile generation
® Runtime testing ® Remote recalls management

Fig. 6.9.3 Remote vehicle diagnostics solution

[XE] PERSONAL : IoT in Healthcare


IoT devices can be used to enable remote health monitoring and emergency
notification systems. These health monitoring devices can range from blood
pressure and heart rate monitors to advanced devices capable of monitoring
specialized implants.
Smart health systems provide health related services using a network, some kind
of connection between intelligent agents. These intelligent agents could be
computing devices, mobile phones, sensors, Fitbit smart bands, surgical devices,
devices that measure your blood chemistry, or devices that measure your
brainwaves. Any of these things could be intelligent agents.
The human actors, patients or healthcare providers for example could be
intelligent agents in this system. The sensors, devices, computers, applications, and
human actors are all intelligent agents that might be connected in the smart health
system.
Smart healthcare is an important research area for Internet of Things, which
employs sensors and other information identifying technologies, wireless and
wired networks to realize large-scale, multi-layer interaction between patients and
medical equipments, medical staff and healthcare institutions.
Some challenges in the healthcare system are as follows :
1. Smarter hospital : Smarter hospital is an important improvement of smart
healthcare system. A natural problem is how to build a smarter hospital for
greatly improving medical services and patient experience.
2. Data integration/realtimeness : How to combine heterogeneous health data
sources in a unified and meaningful way enables the discovery and monitoring
of health data from different sources. It is also important for smart healthcare
to ensure the data realtimeness.

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 6-28 Advanced Techniques in Cloud Computing

3. Medical resource shoriness : There are not enough medical resources for the
population. For example, there are fewer doctors and high-level healthcare
institutions but more patients.
"Low" usage of community health service centers. In contrast with community
health service centers, people prefer the high-level healthcare institutions. This
results in the low usage of community service centers.
Bad health habits : The citizens have some bad health habits that contribute to
poor health, for instance, smoking and no sport.
Lack of information sharing : Hospitals are not sharing enough information.
This leads to the following two problems at least. First, the health information
records of patients cannot be queried. Second, there is lack of medical
cooperation between hospitals.
e The links between the many applications in health monitoring are :
1. Applications require the gathering of data from sensors
2. Applications must support user interfaces and displays
3. Applications require network connectivity for access to infrastructural services
4 . Applications have in-use requirements such as low power, robustness,
durability, accuracy and reliability.
e Connected medical devices and associated IoT technologies will primarily be used
to achieve the following capabilities :
1. Access real time visibility of the patient's condition, his/her activities, context
and physiological parameters
Monitor compliance to prescribed treatment, diet and exercise regimes
3. Provide feedback and cues to patients, family members, doctors and caregivers
in order to implement corrective action
Leverage high performance computing for real time feedback and use
evidence-based medicine for better patient outcome.

EXT] Multiple Choice Questions


Q.1 Which of the following is NOT cloud application features ?

a| Multitenancy b| Elasticity
Cc Homogeneous cloud platform d| On-demand service

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 6-29 Advanced Techniques in Cloud Computing

Q.2 are a set of agreements that are signed between the user and service
providers.

[al Service level agreement [b] Service oriented architecture


Service layer agreement [d] Software level agreement

Q.3 Which of the following is associated with considerable vendor lock-in ?

[a] PaaS [b] TaaS

CaaS [d] SaaS

Q.4 Kubernetes uses to run images and manage containers.

[a] Jungle [b] Docker

AWS [d| None

Q.5 Jungle computing is computing system.

[a] parallel [b] cluster

grid [d| distributed

Q.6 Docker engine is a based application.

[a] client [b] server

client-server [d] all of these

Answer Keys for Multiple Choice Questions :

Q1 c Q2 a Q3 a Q4
Q.5 d Q.6 c

aaa

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing 6-30 Advanced Techniques in Cloud Computing

Notes

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Solved Model Question Paper (In Sem)
Cloud Computing
T.E. (Computer) Semester - VI Elective - II (As Per 2019 Pattern)

Time : 1 Hour] [Maximum Marks : 30

NB: i) Attempt Q.1 or Q.2, Q.3 or Q4.


ii) Neat diagrams must be drawn wherever necessary.
iii) Figures to the right side indicate full marks.
iv) Assume suitable data, if necessary.
Q.1 a) What is cloud computing ? Explain it. [Refer section 1.1] 3)

Explain characteristics of cloud computing. [Refer section 1.2.1] @

Explain various cloud service models. [Refer section 1.6] ®


OR
Explain difference between public and private cloud. [Refer section 1.10.1] 3)

What is Infrastructure as a Service (IaaS) ? List merits and demerits.


[Refer section 1.6.3] 5
Discuss briefly cloud computing reference model. [Refer section 1.8.2] @

What is NAS ? Compare NAS and SAN. [Refer section 2.1] @

Explain data intensive technologies for cloud computing. [Refer section 2.6.5] @

Explain cloud data stores. Disucss data store types. [Refer section 2.4] @
OR
Q4 a) What is cloud data management interface ? [Refer section 2.6.2] 3)

What is cloud file system ? [Refer section 2.3.3] 5)

What is cloud provisioning ? Explain its benefits and characteristics. Also explain its types.
[Refer section 2.2.3] (v4)

M-1)
Cloud Computing M-2 Solved Model Question Papers

Solved Model Question Paper (End Sem)


Cloud Computing
T.E. (Computer) Semester - VI Elective - II (As Per 2019 Pattern)

Time : 25 Hours] [Maximum Marks : 70

NB: i) Attempt Q1 or Q.2, Q.3 or Q4, Q.5 or Q.6, Q.7 or Q.8.


ii) Neat diagrams must be drawn wherever necessary.
iii) Figures to the right side indicate full marks.
iv) Assume suitable data, if necessary.

Q.1 a) What is server virtualization ? Explain requirement of server virtualization. List advantages
and disadvantages of server virtualization. [Refer section 3.3.2] ®

b) Explain full and para virtualization with examples. [Refer sections 3.3.4 and 3.4] (10)
OR
Q.2 a) Explain in brief virtual clusters and resource management. [Refer section 3.6] 5

b) Explain different levels of virtualization implementation with neat diagram. Also give
example of each. [Refer section 3.3] 5]

€) Explain anatomy of cloud infrastructure. [Refer section 3.10.1] 8)

Q.3 a) Write a note on services offered by Amazon ? [Refer section 4.1] ®

b) Explain steps to configure server for EC2 ? [Refer section 4.2] (0)
OR
Q.4 a) Explain the steps to create an Amazon S3 bucket and managing associated objects ?
[Refer section 4.2.2] @®
b) Discuss Google App engine. [Refer section 4.7] ©

Q.5 a) What is risk management ? Explain risk in cloud computing. [Refer section 5.1] (3)

b) Explain cloud security services. [Refer section 5.6] ©)

¢) Discuss secure cloud software requirements. [Refer section 5.8] 6)


OR
Q.6 a) Briefly explain cloud security issues and challenges. [Refer section 5.4] 6)
b) Discuss enterprise-wide risk management in cloud. [Refer section 5.2] ©)

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing M-3 Solved Model Question Papers

c) How secure cloud software testing is performed? Discuss types of cloud testing.
[Refer section 5.9] (3)

Q.7 a) Explain traditional as well as docker deployment workflow, Client server architecture of
Docker ? [Refer section 6.7] ©

b) Describe cloudlets for mobile cloud computing with neat diagram and differentiate between
cloudlets and clouds. [Refer section 6.2] (3)
OR
Q.8 a) What is energy aware cloud computing ? Explain in detail. [Refer section 6.5] ®)

b) What is automatic cloud computing ? Discuss briefly CometCloud. [Refer section 6.3] (9)

aaa

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge


Cloud Computing M-4 Solved Model Question Papers

TEXT BOOKS FOR TE. (COMP) SEM VI


| Compulsory
Compulsory Subjects
Subjects _
1. Web Tech nology (A. A. Puntambekar)

2. Data Science and Big Data Analytics (. a. bhotre, or. kalpana v. Metre)
3. Artificial Intelligence {Anamitra Deshmukh-Nimbalkar, Dr. Vaishali P. Vikhe)

Elective Subjects

4. Information Security (. a. photre, br. swati Nikam)


5. Augmented and Virtual Reality (or. ninad More, sunita patil)
6. Cloud Computing (. a. bhotre)
7. Software Modeling and Architecture (a. a. puntambekar)
\. J
BREE

A Guide for Engineering Students

PAPER SOLUTIONS
m

© Covers Entire Syllabus e Question Answer Format ¢ Exact Answers & Solutions
© Important Points to Remember e Important Formulae
For All © Chapterwise Solved University Questions Last 10 Years Solved Papers
Branches
... Available at all Leading Booksellers ...

TECHNICAL PUBLICATIONS® - an up-thrust for knowledge

You might also like