Cloud Computing
Cloud Computing
CLouD COMPUTING
Iresh A. Dhotre
M.E. (Information Technology)
Ex-Faculty, Sinhgad College of Engineering,
Pune.
= "TECHNICAL
I~
N=
~~ P
~~ |
SINCE 1993 An Up-Thrust for Knowledge
U]
CLOUD COMPUTING
Published by :
="TECHNICAL Amit Residency, Office No.1, 412, Shaniwar Peth,
S PUBLICATIONS Pune - 411030, M.S. INDIA, Ph.: +91-020-24495496/97
sce 255 An Up-Thrust for Knowledge Email : [email protected] Website : www.technicalpublications.org
Printer :
Yogiraj Printers & Binders
Sr.No. 10/1A,
Ghule Industrial Estate, Nanded Village Road,
Tal. - Haveli, Dist. - Pune - 411041.
ISBN 978-93-91567-40-8
939115674 SPPU 19
The book uses plain, lucid language to explain fundamentals of this subject. The book
provides logical method of explaining various complicated concepts and stepwise methods
to explain the important topics. Each chapter is well supported with necessary illustrations,
practical examples and solved problems. All the chapters in the book are arranged in a
proper sequence that permits each topic to build upon earlier studies. All care has been
taken to make students comfortable in understanding the basic concepts of the subject.
Representative questions have been added at the end of each section to help the
students in picking important points from that section.
The book not only covers the entire scope of the subject but explains the philosophy of
the subject. This makes the understanding of this subject more clear and makes it more
interesting. The book will be very useful not only to the students but also to the subject
teachers. The students have to omit nothing and possibly have to cover nothing more.
[ wish to express my profound thanks to all those who helped in making this book a
reality. Much needed moral support and encouragement is provided on numerous
occasions by my whole family. | wish to thank the Publisher and the entire team of
Technical Publications who have taken immense pain to get this book in time with quality
printing.
Any suggestion for the improvement of the book will be acknowledged and well
appreciated.
Author
D, A, Dhotre
Dedicated to God.
(iid)
SYLLABUS
Cloud Computing - 310254(C)
Credit : Examination Scheme :
(iv)
UnitV Security in Cloud Computing
Risks in Cloud Computing : Risk Management, Enterprise-Wide Risk Management, Types of Risks
in Cloud Computing. Data Security in Cloud : Security Issues, Challenges, advantages,
Disadvantages, Cloud Digital persona and Data security, Content Level Security. Cloud Security
Services : Confidentiality, Integrity and Availability, Security Authorization Challenges in the Cloud,
Secure Cloud Software Requirements, Secure Cloud Software Testing. (Chapter - 5)
™
TABLE OF CONTENTS
1.7 Cloud Computing Logical Architecture .......cccce eevee ccee ccc cceeeceeeeceee 1-17
1.7.1 Cloud Computing Service Architecture.......ccooveeiieeee eine 1-19
1.9 Cloud System Archit@CtUIE .....cooccieei ieee cee ec ee es rree ee eeesree ee sreeeeaee 1-24
1.9.1 Single Cloud Site ArchiteCtUres......coceevieeiiireeieee eee 1-25
1.11 Multiple Choice Questions With ANSWErS .......ccccevcieer icine eves 1-30
2.1.4 Comparison between NAS and SAN .......ccorr ieee eee 2-9
2.1.5 Comparison between DAS, NAS and SAN .......coccoviirinienes sees 2-10
2.2 Data Storage Management. ...c.eei ieee ccciieeie cece e essere reese esr e eee 2-10
2.2.1 Data Storage Management TOOIS .......c.coeveireeiieneneee
eee eee 2-10
0 J TLR V3 Yo SR 2-12
2.3.1 FAT FIle SYSEEM coon eae sae 2-13
2.4 CloUd Data StOMES.... ii ieieiueieieeeiee site ree erica ete setae see e sues ete eseeesaee sabes see e anne enns 2-15
2.4.1 Distributed Data Store ......cceveee eerie 2-16
2.4.2.1 BIgTABI cc. iiiete ste staat ste eae anata ata een 2-16
3.5 Storage Virtualization ......cee cece ccc ceeserre eer ee ee ee ee ee ee eee 3-15
3.5.1 Network Virtualization .........ceeeeeeeeieeeeeireeeee cee 3-18
3.6 Virtual ClUSLEIING eee iceeeciee ects ects estes stresses estes ea reese re ese rare essere ee enes 3-19
3.7 Virtualization Application......cceeeiciir ices ccee ec eeeeseeeeeeeeeee 3-20
(vii)
3.10 Virtualization and Cloud COMPUTING ......coecciereiiier reciever sees eens 3-25
3.10.1 Anatomy of Cloud INfrastructure........coceeveeeeiie ieee 3-25
Unit - IV
4.2 Elastic Cloud COMPUEING c.cuveveiiieie ices csi crie eee e esses erase esas senna e anes 4-4
4.2.1 Configuring Amazon EC2 LINUX INStANCES ....coverierierer cies eee 4-6
4.3 Amazon STOrage SYSTEM uve essere essere reese errr essere eee eens 4-13
4.4 Amazon Database SErvICeS ......ccuiii iii ieee erie eee ese eee eee eee 4-14
4.41 DYNAMO DB eee eee 4-15
4.6.4 Business and Consumer Applications : CRM and ERP ........ccocceeievrenennnne 4-23
(ix)
Chapter - 5 Security in Cloud Computing (5-1) to (5-14)
5.1 Risks in Cloud COMPULING....cieiiiiireiiier eters scieressee essere essere seers se ree essen snes 5-2
5.1.1 Risk Management ........ooieeieiin ieee cee eee ee eee 5-2
Unit - VI
(SA Vo]
o 11 T= [TH oR 6-3
6.2.1 Advantages and Disadvantages .........cceeeevvierieireniiesee eee 6-5
6.5.4 Key Issue Related to Cloud Computing Energy Efficiency ........ccooveveenenne 6-15
6.8 INTroduction t0 DEVOPS ...ueeieicieee cesses crtre ee ertte essere erates ses eennre es senre ee enes 6-23
6.9 10T and Cloud CONVEIENCE. ......eii ieee ecieeeeciee essere essere eeenre ese snses assess snnreeeenes 6-23
6.9.1 The Cloud and 10T in yoUr HOME ....cc.coiiiiiiirieiieee sees 6-23
6.9.2 The IOT and Cloud in your Automobile .........cocovvvirieiini eee 6-25
(xii)
UNIT - 1
Syllabus
Importance of Cloud Computing, Characteristics, Pros and Cons of Cloud Computing, Migrating
into the Cloud, Seven-step model of migration into a Cloud, Trends in Computing. Cloud Service
Models : SaaS, PaaS, laa, Storage. Cloud Architecture : Cloud Computing Logical Architecture,
Developing Holistic Cloud Computing Reference Model, Cloud System Architecture, Cloud
Deployment Models.
Contents
1.1 Overview of Cloud Computing ............. March-19, ------------- Marks 5
1.2 Importance of Cloud Computing. . .......... March-19, ------------- Marks 5
1.3 Migrating into the Cloud
14 Seven-step Model of Migration into a Cloud
1.5 Trends in Computing
1.6 Cloud Service Models .................. March-19, ------------- Marks 4
1.7 Cloud Computing Logical Architecture
1.8 Developing Holistic Cloud Computing Reference Model
1.9 Cloud System Architecture
1.10 Cloud Deployment Models . ............... March-19, ------------- Marks 6
1.11 Multiple Choice Questions
(1-1)
Cloud Computing 1-2 Introduction to Cloud Computing
¢ Cloud computing is a general term used to describe a new class of network based
computing that takes place over the Internet, basically a step on from utility
computing.
e In other words, this is a collection/group of integrated and networked hardware,
software and Internet infrastructure (called a platform).
e Fig. 1.1.1 shows cloud symbol. It
denotes cloud boundary.
e Using the Internet for
communication and transport
provides hardware, software and
networking services to clients.
e These platforms hide the complexity
and details of the underlying
infrastructure from users and
applications by providing very
simple graphical interface or APL Fig. 1.1.1 Cloud symbol
e In addition, the platform provides on demand services that are always on
anywhere, anytime and anyplace. Pay for use and as needed, elastic.
e The hardware and software services are available to the general public, enterprises,
corporations and business markets.
\\
Wy
My
\\
Data centers
Client computers
\\
\\
Distributed servers
1. What is cloud computing ? What are the components of cloud computing ? Explain with neat
diagram. SPPU : March-19, In Sem, Marks 5
Cloud computing offers the media streaming companies a cost effective, flexible,
secure and scalable solution for setting up their work-flow. This serves to meet the
media streaming companies requirements for both video on demand and live
streaming services.
During a live broadcast of India v/s Pakistan match, the traffic is easily 10-20
times that on normal days. Provisioning for servers and other infrastructure would
end up creating a large overlay of upfront cost for infrastructure which otherwise
would not be utilised in normal days. The cloud with its pay per use strategy
offers economies of scale with effective costing strategies at a very high rate of
reliability and security.
4. Instant software updates : When you access a web-based application, you get the
latest version - without needing to pay for or download an upgrade.
5. Improved document format compatibility : You do not have to worry about the
documents you create on your machine being compatible with other user's
applications or operating systems.
6. Unlimited storage capacity : Cloud computing offers virtually limitless storage.
7. Increased data reliability : Unlike desktop computing, in which if a hard disk
crashes and destroy all your valuable data, a computer crashing in the cloud
should not affect the storage of your data.
8. Universal document access : All your documents are instantly available from
wherever you are.
9. Latest version availability : The cloud always hosts the latest version of your
documents; as long as you are connected, you are not in danger of having an
outdated version.
10. Easier group collaboration : Sharing documents leads directly to better
collaboration.
11. Device independence : Move to a portable device and your applications and
documents are still available.
cloud can help them adapt to diverse user groups with a diverse assortment of
devices.
4. Masked complexity enables access to services, no matter how intricate the
technology they're built on.
5. With context-driven variability, "intelligent assistants” are possible. "Because of its
expanded computing power and capacity, cloud can store information about user
preferences, which can enable product or service customization,” the report states.
6. Ecosystem connectivity enables information exchange across business partners.
architecture layers. You may employ these across hybrid and multi-cloud
environments based on business needs and best practices.
¢ Cloud migration strategy should focus on four broad areas :
a) Security that determines application and data access controls, data security at
rest and in transit and backup and data recovery.
b) Governance and compliance that makes sure you meet regulatory compliance
and all aspects of data security to go beyond the limits of shared security
models of cloud services providers.
¢) Cost management that looks at immediate and long-term cost savings that
guide where and how you migrate, manage, and monitor applications and
workloads.
d) Accessibility, resilience, and scalability that are driven by customer and
workforce UX needs, service enhancements and innovation.
e Benefits of migrating to the cloud include :
a) Increased agility and flexibility
b) Ability to innovate faster
c) Easing of increasing resource demands
d) Better managing of increased customer expectations
e) Reduction in costs
f) Deliver immediate business results
g) Simplify IT
h) Shift to everything as-a-service
i) Better consumption management
j) Cloud scalability
k) Improved performance
Cloud migration process :
e Cloud migration process is divided into three phases : plan, execute (run) and
monitor
1. Plan : Cloud migration requires solid planning to be successful. It define following
parameters :
a. Identify business objective
b. Identify key business drivers
c. Get executive sponsorship
d. Providing full visibility into your on-premise environment, including all system
dependencies.
Cloud compatible
Current cost
Nowe
These assessments are about the cost of migration as well as about the ROI that
can be achieved in the case of production version.
Step 2 : An isolating all systemic and environmental dependencies of the
enterprise application components within the captive data center.
Step 3 : Generating the mapping constructs between what shall possibly remain in
the local captive data center and what goes onto the cloud.
Step 4 : Substantial part of the enterprise application needs to be rearchitected,
redesigned, and reimplemented on the cloud.
Step 5 : We leverage the intrinsic features of the cloud computing service to
augment our enterprise application in its own small ways.
Step 6 : We validate and test the new form of the enterprise application with an
extensive test suite that comprises testing the components of the enterprise
application on the cloud as well.
Step 7 : Test results could be positive or mixed. In the latter case, we iterate and
optimize as appropriate. After several such optimizing iterations, the migration is
deemed successful.
1. IT maturity analysis
The technology maturity analysis, which compares the technologies against each
other in terms of their adoption rates and current investment rates.
Major trends that emerged in cloud computing technology are small and medium
business, hybrid cloud model etc.
2. Technology Trends to Watch
Virtualization is technology that lets user create useful IT services using resources
that are traditionally bound to hardware.
Virtualization is composed of server, storage, desktop, application, infrastructure
and network. It is compatible with today's technology.
Data growth : Day by day, data may goes on increasing and organization can not
provide storage. So cloud is solution for storing and processing of growing data.
Green IT : The term green IT means nothing other than the use of information
technology in a manner that preserves resources and protects the environment.
Cloud computing, the provision of IT infrastructure via the Internet, plays an
important role in green IT. Cloud allows the replacement of poorly utilized data
centers with highly optimized server environments. Besides reducing the energy
costs, it also minimizes the time required to maintain a company's proprietary
systems.
The service models build on one another and define what a vendor must manage
and what the client's responsibility is.
Service models : This consists of the particular types of services that you can
access on a cloud computing platform.
Cloud service is any service made available to users on demand via the Internet
from a cloud computing provider's servers as opposed to being provided from a
company's OWI on-premises servers.
Cloud services are designed to provide easy, scalable access to applications,
resources and services, and are fully managed by a cloud services provider.
A cloud service can exist as a simple web-based software program with a technical
interface invoked via the use of a messaging protocol, or as a remote access point
for administrative tools or larger environments and other IT resources.
The organization that
provides cloud-based IT
Cloud Clients
resources is the cloud
Web browser, mobile app, thin client, terminal
provider. Cloud providers emulator, ...
normally own the IT
write some points of saas, paas
resources for lease by cloud | and Iaas also
consumers, and could also .
resell IT resources leased 2 SaaS
from other providers. 2 CRM, Email, virtual desktop, communication,
. 3 games, ...
Cloud computing, often
. £
described as a stack, has a = PaaS
broad range of services built ES Execution runtime, database, web server,
a development tools, ...
on top of one another under
the name cloud. laaS
structure
Infra
the above fig is the cloud computing stack which showing the service models of cloud computing
Cloud Computing 1-12 Introduction to Cloud Computing
3. IaaS is the hardware and software that powers it all - servers, storage, networks,
operating systems.
Users can access the application on the internet without installing
the application on local storage thus because of it saves the
EEX] software as a Service (SaaS) computer hardware also.
| Application
: ~f—
Client
H
i
Service provider
offering SaaS
Client
Characteristics of Saa$ :
Benefits of Saa$ :
Easier collaboration
It support automated update and patch management services.
¢ The customer still does not manage the framework, network, servers or operating
system, but has control over deployed applications and sometimes over the
hosting environment itself.
Some examples of Platform as a Service include Google's App Engine or Force.com
PaaS consists of following components :
1. Browser based development studio
2. Pay contrary to billing
3. Management and supervising tools
4. Seamless deployment to host run time environment.
Characteristics of Paa$ :
1. It support multi-tenant architecture.
2. It support for development of group collaboration.
3. PaaS systems can be deployed as public cloud services or as private cloud
services.
Provision of runtime environments. Typically each runtime environment
supports either one or a small set of programming languages and frameworks
Support for custom applications. Support for the development, deployment and
operation of custom applications.
Preconfigured capabilities. Many PaaS systems are characterized by capabilities
that are preconfigured by the provider, with a minimum of configuration
available to developers and customer operations staff.
Support for porting existing applications. While many PaaS systems are
primarily designed to support "born on the cloud” applications.
Security is an important characteristic in PaaS. It needs to provide
authentication and authorization to differentiate the access rights of different
users
Benefits of Paas :
1. Scalability including rapid allocation and deallocation of resources with a
pay-as-you-use model
Reduced capital expenditure
WN
IaaS gives the storage room likeness to the in-house datacenter stood out from
various organizations sorts.
Center datacenter framework segments are capacity, servers (registering units), the
system itself, and administration apparatuses for foundation upkeep and checking.
Each of these parts has made a different market specialty. While some little
organizations have practical experience in just a single of these IaaS cloud
specialties, vast cloud suppliers like Amazon or Right Scale have offerings over all
Taa$S territories.
Fig. 1.6.3 shows IaaS.
Physical server
Enterprise
Data storage
device
Computer
User
It offers the hardware so that your organization can put whatever they want onto
it. Rather than purchase servers, software, racks, and having to pay for the
datacenter space for them, the service provider rents those resources :
1. Server space 2. Network equipment
3. Memory 4. CPU cycles 5. Storage space
Again, the customer is not managing cloud infrastructure, but in this case, the
customer does control operating systems, deployed applications, storage, and
sometimes-certain networking components
Examples : Amazon EC2, Rackspace Mosso, GoGrid
3. Shared virtual server : The customer can access a virtual server on a device
that may be shared with other customers.
Advantages of laa$S :
1. Elimination of an expensive and staff-intensive data center
Ease of hardware scalability
Nn
TaaS gives users automated PaaS provides a framework SaaS makes applications
and scalable environments for quickly developing and available through the internet.
deploying applications
Amazon Web Services, for Google Cloud Platform SaaS applications such as
example, offers IaaS through provides another PaaS option Gmail, Dropbox, Salesforce, or
the Elastic Compute Cloud, or in App Engine Netflix
EC2
Major IaaS providers include Examples of PaaS services are email services such as Gmail
Amazon Web Services, Heroku and Google App and Hotmail are examples of
Microsoft Azure, and Google Engine. cloud-based SaaS services.
Compute Engine.
IaaS services are available on ~~ PaaS solutions are available SaaS services are usually
a pay-for-what-you-use model with a pay-as-you-go pricing available with a pay-as-you-go
model. pricing model
7] I
x 7 [4 \r __r x
|
Service request examiner and I
admission control !
- Customer - driven service management |
- Computation risk management !
- Automatic resource management
SLA i
Resource !
Allocator = - \
Pricing Accounting 1
||
i
VM Service request |
Monitor Dispatcher monitor !
|
|
|
|
i|
Virtual !
Machines I
(VMs)
|
I
|
|
|
i|
Physical !
Machines
i
I
|
|
1. Users/Brokers : They submit their service requests from anywhere in the world
to the cloud.
2. SLA resource allocator : It is a kind of interface between users and cloud
service provider which enable the SLA-oriented resource management.
3. Service request examiner and admission control : It interprets the submitted
request for QoS requirements before determining whether to accept or reject
the request. Based on resource availability in the cloud and other parameters
decide.
4, Pricing : It is in charge of billing based on the resource utilization and some
factors. Some factors are request time, type etc.
5. Accounting : Maintains the actual usage of resources by request so that the
final cost can be charged to the users.
6. VM monitor : Keeps tracks on the availability of VMs and their resources.
7. Dispatcher : The dispatcher mechanism starts the execution of admitted
requests on allocated VMs.
8. Service request monitor : The request monitor mechanism keeps track on
execution of request in order to be in tune with SLA.
Cloud components :
¢ Cloud computing solutions are made up of several elements.
1. Clients : Mobile, terminals or regular computers.
2. Benefits : Lower hardware costs, lower IT costs, security, data security, less
power consumption, ease of repair or replacement, less noise.
3. Data centers : Collection of servers where the application to subscribe is
housed. It could be a large room in the basement of your building or a room
full of servers on the other side of the world
4. Virtualizing servers : Software can be installed allowing multiple instances of
virtual servers to be used and a dozen virtual servers can run on one physical
server.
5. Distributed servers : Servers don't all have to be housed in the same location.
It can be in geographically disparate locations. If something were to happen at
one site, causing a failure, the service would still be accessed through another
site. If the cloud needs more hardware, they can add them at another site.
Cloud
Cloud platform development Platform as a service
environment
Virtual Infrastructure as a
Cloud infrastructure infrastructure service
manager
¢ Cloud computing is one of the next significant stage in the Internet's evolution,
providing the means through which everything from computing power to
computing infrastructure, applications, business processes to personal collaboration
can be delivered to the user as a service wherever and whenever they need.
e The "cloud" in cloud computing can be defined as the set of hardware, networks,
storage, services and interfaces that combine to deliver aspects of computing as a
service.
e Consumer cloud computing services has been well established ever since
mainstream Internet. Known examples are Webmail services and social networking
platforms.
e However the adoption of cloud computing within the enterprise sector has been
slow. This slow uptake in cloud services that promises so much has been
primarily influenced by the numerous security risks, concerns and challenges
posed within such an environment.
¢ Governance, risk and compliance factors of cloud services need to be fully
assessed and evaluated by organizations to provide informed judgments. Data and
information lifecycle, source and origination, transfer, destination, validation and
deletion all need to be understood.
e Despite significant attention and substantial efforts both in industry and academia,
cloud computing has not yet reached its full potential. Commonly stated obstacles
for cloud adoption include confusion about terminology with multiple delivery
models (SaaS, PaaS, IaaS) and deployment cloud scenarios.
Software as a Service
user2 | &&——p»
Platform as a Service
UserN | {>
Infrastructure as a Service
eo Virtualized servers
e Storage and networking
® Examples : Amazon EC2, S3, vCloud
Software as a Service
e SaaS provider dispose the applied software unified on their server, the user can
subscribe applied software service from the manufacturer through Internet.
e Provider supply software pattern through Browser, and charge according to the
quantity of software and using time. The advantage of this kind of service pattern
is that the provider maintains and manages software, supplies the hardware
facilities, the users can use software everywhere when they own the terminal
which can log in Internet.
Platform as a Service (PaaS) :
e PaaS takes develop environment as a service to supply. This layer provides a
platform for creating applications.
e PaaS solutions are essentially development platforms for which the development
tool itself is hosted in the cloud and accessed through a browser.
e With PaaS, developers can build web applications without installing any tools on
their computers and then deploy those applications without any specialized
systems administration skills.
¢ Google App Engine is the representative product through their server and
Internet.
Single Server
e A single server means only one server, whether physical or virtual, containing one
web server, database, and application.
eo LAMP stack is an example of single server. This structure is not widely utilized
due to associated security risks.
rd AN
Load balancer 1 Load balancer 2
Application 1 Application 2
Snapshot
backups
DNS
Application 1 Application 2
= amd
Replication Slave <a»
Snapshot
backups
e Cloud deployment models are refers to the location and management of the
cloud’s infrastructure.
¢ Deployment models are defined by the ownership and control of architectural
design and the degree of available customization. Cloud deployment models are
private public and community clouds.
e Fig. 1.10.1 shows cloud deployment model.
1. Public cloud :
¢ The cloud infrastructure is made available to the general public or a large industry
group and is owned by an organization selling cloud services.
e Public cloud is a huge data centre that offers the same services to all its users. The
services are accessible for everyone and much used for the consumer segment.
e Examples of public services are Facebook, Google and LinkedIn.
Private cloud
/_aa8
“2
(a) Private cloud (b) Public cloud
Fig. 1.10.1
e Public cloud benefits :
a) Low investment hurdle : Pay for what user use.
b) Good test/development environment for applications that scale to many
servers.
Public cloud risks :
a) Security concerns : Multi-tenancy and transfers over the Internet.
b) IT organization may react negatively to loss of control over data center
function.
2. Private cloud :
The cloud infrastructure is operated solely for a single organization. It may be
managed by the organization or a third party and may exist on-premises or
off-premises.
Private cloud benefits :
a) Fewer security concerns as existing data center security stays in place.
b) IT organization retains control over data center.
Private cloud risks :
a) High investment hurdle in private cloud implementation, along with purchases
of new hardware and software.
b) New operational processes are required; old processes not all suitable for
private cloud.
Organization 1 Organization 2
3. Community cloud :
e The cloud infrastructure is shared by several organizations and supports a specific
community that has shared concerns (e.g. mission, security requirements, policy, or
compliance considerations). It may be managed by the organizations or a third
party and may exist on-premises or off-premises.
4. Hybrid cloud :
eo The cloud infrastructure is a composition of two or more clouds (private,
community or public) that remain unique entities but are bound together by
standardized or proprietary technology that enables data and application
portability (e.g., cloud bursting for load-balancing between clouds).
e Hybrid cloud benefits :
a) Operational flexibility : Run mission critical on private cloud, dev/test on
public cloud
b) Scalability : Run peak and bursty workloads on the public cloud
¢ Hybrid cloud risks :
a) Hybrid clouds are still being developed; not many in real use
b) Control of security between private and public clouds, some of same concerns
as in public cloud
Public cloud infrastructure is offered via web Private cloud infrastructure is dedicated to a
applications and also as web services over single organization.
Internet to the public.
Azure, Amazon Web Services, Google App An example of the Private Cloud is NIRIX’s
Engine and Force.com are a few examples of one Server with dedicated servers
public clouds
Cloud computing makes the long-held dream of utility computing possible with
a pay-as-you-go, infinitely scalable, universally available system.
ld All of the mentioned.
Q.9 is the entry point into the cloud for user and administrators.
[a] Cloud Manager [b] Group Manager
Instance Manager [d] VM Manager
Q.10 The services that provides utility may directly correlate with their ?
Q.11 The promise of has raised the IT expectations of small and medium
enterprises beyond measures.
Q.16 In what type of cluster, the nodes are closely packaged in one or more racks sitting
in a room, and the nodes are not attached to peripherals ?
[a] 4 [b] 5
[c] 6 ld] 7
Q.20 _____ is the process of transferring data between storage types, formats or systems.
Q1 c Q.2 b Q3 c Q4 a
Q5 c Q.6 c Q7 a Qs b
aaa
Syllabus
Data Storage : Introduction to Enterprise Data Storage, Direct Attached Storage, Storage Area
Network, Network Attached Storage, Data Storage Management, File System, Cloud Data Stores,
Using Grids for Data Storage. Cloud Storage : Data Management, Provisioning Cloud storage,
Data Intensive Technologies for Cloud Computing. Cloud Storage from LANs to WANs : Cloud
Characteristics, Distributed Data Storage
Contents
2.1 Introduction to Enterprise Data Storage
2.2 Data Storage Management
2.3 File System LL... March-19,20, ----------- Marks 6
2.4 Cloud Data Stores ~~ .................. Dec.-16,17, March-19, - - - - Marks 5
2.5 Using Grids for Data Storage
2.6 Cloud Storage
2.7 Cloud Storage from LANs to WANs
2.8 Multiple Choice Questions
Direct - Attached Storage (DAS) is hard disk drives or solid - state drives
connected directly inside or outside (in a storage enclosure) to a single computer
or server that cannot be accessed by other computers or servers. DAS is not
networked through Ethernet or FC switches.
The storage device may include one or more drives built into a server and with an
appropriate host bus adapter, may be configured as a RAID array.
Examples of DAS include hard drives, optical disc drives and storage on external
drives.
One advantage of DAS storage is its low initial cost. A key disadvantage of DAS
storage is its limited scalability. A Host Bus Adaptor can only support a limited
number of drives.
Direct attached storage tends to be less expensive than NAS and SAN
implementations, but lacks scalability and flexibility.
DAS are of two types : Internal DAS and external DAS.
1. Internal DAS : Storage device is internally connected to the host. The
connection between host and storage device is serial or parallel bus.
2, External DAS : Server connects directly to the external storage device.
Advantages of DAS :
1. DAS is ideal for localized file sharing in environments with a single server or a
few servers.
DAS devices can offer block-level access or file-level access.
DAS also offers ease of management and administration.
ISA
CR
Server
Storage
area
network
Disks BE TE
e SANs are primarily used to enhance storage devices, such as disk arrays, tape
libraries and accessible to servers so that the devices appear like locally attached
devices to the operating system
e SAN systems provide a simple, untyped, fixed-size (block), memory-like interface
for manipulating nonvolatile magnetic media. SAN is used when multiple servers
need access to the same data (Example : Databases).
e A SAN is a high - speed network of storage devices, separate from the local area
network that also connects those storage devices with servers. It provides block -
level storage that can be accessed by the applications running on any and all
servers connected to the SAN. SAN storage devices can include tape libraries and
more commonly, disk - based devices, like RAID hardware.
e A storage area network is designed for managing very large amounts of network
storage. For large organizations with many servers that need access to the same
data, SAN offers better performance and flexibility than DAS or NAS.
e Storage area network is defined as a set of interconnected devices and servers that
are connected to a common communication and data transfer infrastructure such
as fiber channel. The purpose of the SAN is to allow multiple servers access to a
pool of storage in which any server can potentially access any storage unit.
e A SAN is a network designed to transfer data from servers to targets and it is
alternative to directly attached target architecture or to a DAS architecture, where
the storage is connected to the servers on general purpose networks.
eo A SAN can be considered as an extended and shared storage bus within a data
center, consisting of various storage devices and specific interfaces (e.g. fibre
channel) rather than the Ethernet. SANs provide high - bandwidth block storage
access over long distance via extended fiber channel links.
SAN features :
1. Network resource used exclusively for the storage.
2. SCSI protocol for end to end communication : Minimum impact for DAS to
SAN migration.
Access to data through logical blocks and not to file.
Uk Ww
Scalability like a LAN/WAN, it can use a variety of technologies. This allows easy
relocation of backup data, operations, file migration and data replication between
systems.
Advantages of SAN :
SAN support large and heterogeneous block data transfer
ok whe
Resilience to failure.
Server 1 Server 2
\ /
/
IP network
Storage arrays
2. Accessing file system : File system must be mounted before it can be used.
During boot process, operating system mounts the local file system. To mount
any file system, two pieces of information must be specified :
a. A means of uniquely identifying the desired disk drive and partition, such
as device file name, file system label.
¢ An integrated NAS solution is available with low - end device to high end
solution. They are connected to the local area network and communicate with file
systems through Trans - Mission Control Protocol / Internet Protocol (TCP / IP).
2. Gateway NAS :
eo It is NAS device with the internal processor but without integrated storage.
Instead, the NAS device connects to storage by direct attachment or by a SAN.
e In a NAS gateway, the proprietary RAID coniroller is replaced with an open,
ANSI-standard SAN interface, usually a fibre channel host bus adapter. This
allows the NAS gateway to use shared, nonproprietary storage, which provides
the flexibility for an effective and scalable storage strategy.
e When deploying NAS gateways, it is important to select the right shared storage
platform to house your organization's data.
¢ Fig. 2.1.4 shows NSA gateway.
FC
Integrated network switch Fibre
attached storage channel
SAN
1. Integrated NAS uses fibre channel Gateway NAS uses both fibre channel
disks only. and SATA disks.
2. Integrated NAS manages storage Gateway NAS relies on storage array
devices through NAS software. management software to manage
storage devices.
3. It uses specialized operating systems. It uses generic operating systems.
4. Integrated NAS allows both file and Gateway NAS allows only block level
block level access to the clients. access to the clients.
e Benefits of NAS :
1. NAS support NFS and CIFS file system.
Entry level NAS systems are quite inexpensive.
NAS ensure security and user authentication.
ISA
1 Machine connected with LAN may Server class devices that are equipped
utilize NFS, CIES or HTTP protocol to with SCSI and fibre channel adapters
connect to a NAS connect to a SAN.
2. File system is managed by the NAS The SAN servers manage the file system
head unit
3. Backups and mirrors are generated on Backups and mirrors require a block by
files, not blocks block copy operation.
4. A NAS identifies the data by file name SAN addresses the data by logical block
and byte offset, transfers file data or numbers and transfers the data in disk
metadata blocks.
5. NAS uses TCP/IP networks SAN uses fibre channel
¢ Storage monitoring software helps user to keep track of storage capacity and
assigning storage to various applications accordingly. It saves time and user
efficiency. Regular monitoring of the physical health, failures, and outages can
help manage storage devices.
¢ Configuration, migration, provisioning, storage monitoring and reporting are the
management level tasks. Storage Resource Management (SRM) tools include
configuration tools, provisioning tools and measurement tools.
e Storage management tools include capacity management, monitoring, cloud
storage management and other devices and applications to manage SAN and NAS
systems. Storage management also includes backup and secondary storage tools
such as snapshots and copy data management.
1. Configuration tools : It handles the set-up of storage resources. These tools
help to organize and manage RAID devices.
2. Provisioning tools : It defines and control access to storage resources for
preventing a network user from being able to use any other user's storage.
3. Measurement tools : Analyse performance based on behavioural information
about a storage device.
Characteristics of provisioning :
1. Missing and incorrect information is resolved.
2. Product being deliver to customer before provisioning.
The characteristics of a provision are that it is a liability where there is uncertainty
as to either the timing of settlement or the amount to be settled.
When measuring a provision, the amount to be recognized should be the best
estimate of the consideration required to settle the present obligation at the end of
the reporting period.
The fact that it is difficult to measure a provision and that estimates have to be
used does not mean that the provision is not reliably measurable.
Types of cloud provisioning are as follows :
1. Dynamic provisioning : Cloud resources are deployed flexibly to match a
customer's fluctuating demands. The deployments typically scale up to
accommodate spikes in usage and scale down when demands decrease. The
customer is billed on a pay-per-use basis.
2. Self - provisioning : With user self-provisioning, also called cloud self-service,
the customer buys resources from the cloud provider through a web interface
or portal. This usually involves creating a user account and paying for
resources with a credit card.
3. Advance provisioning : Customer undertake contract with the provider for the
required services.
EXE] NTFS
NT File System (NTFS), which is also sometimes called the new technology file
system, is a process that the Windows NT operating system uses for storing,
organizing and finding files on a hard disk efficiently.
Formatting a volume with the NTFS file system results in the creation of several
system files and the Master File Table (MFT), which contains information about all
the files and folders on the NTFS volume.
NTES file system supports larger file sizes and hard drives and is more secure
than FAT. In NTFS, the size of cluster ranges from 512 bytes to 64 kilobytes.
NTES file system can help you to compress files and folders or the whole NTFS
partition to save disk space so you can make best use of it.
Cloud file storage is most appropriate for unstructured data or semi - structured
data, such as documents, spreadsheets, presentations and other file - based data.
Cloud file storage is a storage service that is delivered over the Internet, billed on
a pay-per-use basis and has an architecture based on common file level protocols
such as Server Message Block (SMB), Common Internet File System (CIFS) and
Network File System (NFS).
Cloud File system should be scalable enough to adopt large organizations file
systems under different workloads with good performance requirements.
Cloud file storage is a method for storing data in the cloud that provides servers
and applications access to data through shared file systems. This compatibility
makes cloud file storage ideal for workloads that rely on shared file systems and
provides simple integration without code changes.
A file system in the cloud is a hierarchical storage system that provides shared
access to file data. Users can create, delete, modify, read, and write files and can
organize them logically in directory trees for intuitive access.
Cloud file storage can easily expand with multiple servers as per
Benefits of Cloud File Storage : demand thus it can used for large oragnizations also
Storing file data in the cloud delivers advantages in three key areas :
1. Scalability : Although not every cloud file storage solution can scale,
leveraging all the capabilities of the cloud, the most advanced solutions
provide the ability to start with the capacity and performance you need today
and grow your capacity as needed. No more over provisioning to try and
anticipate future needs.
2. Interoperability : Many existing applications require integration with shared
file services that follow existing file system semantics. Cloud file storage
solutions offer a distinct advantage as there is no new code to write to have
secure, shared file access.
local storage
3. Budget and Resources : Operating file services on-premises requires budget for
hardware, ongoing maintenance, power, cooling, and physical space. Cloud file
storage enables organizations to redeploy technical resources to other projects
that bring more value to the business.
GFS is elastic and cost efficient. It is highly secure. It can be mounted on server,
client or access files via web page.
EX¥X] BigTable
Bigtable is a distributed storage system that is used for managing and storing
structured data at Google.
Bigtable is designed to reliably scale to petabytes of data and thousands of
machine. Bigtable has multiple goals like applicability, high availability, scalability,
high performance.
It is used by approximate sixty Google project or product like Google Analytic,
Google Finance, Personalized search, Writely and Google Earth.
Bigtable is built on Google file system for storing the data for scheduling large
scale data processing. It stored data in form of rows, columns and timestamp that
means it maps with arbitrary string value like row key and colmn key as well as
timestamp.
Features :
4. Bigtable use a highly available and persistent distributed lock services called
Chubby.
5. Bigtable supports single - row transactions.
The map is indexed by a row key, column key and a timestamp ; each value in
the map is an un - interpreted array of bytes.
1. Row key :
The row keys in a table are arbitary strings. Every read or write of data under a
single row key is atomic.
Bigtable maintains data in lexicographic order by row key. The row range for a
table is dynamically parttioned.
Each row range is called a tablet, which is the unit of distribution and load
balancing.
2. Column key :
Column keys are grouped into sets called column families, which form the basic
unit of access control. All data stored in a column family is usually of the same
type.
A column family must be created before data can be stored under any column key
in that family; after a family has been created, any column key within the family
can be used.
XX] Dynamo
Dynamo is propriety key value structured storage system. It can act as database
and also distributed hash table.
Dynamo dynamically partitions a set of keys over a set of storage nodes
It is most powerful relational database available in WWW. Relational databases
have been used a lot in retail sites, to make visitors browse and search for product
easily.
Dynamo does not support replication.
Dynamo is used to manage the state of services that have very high reliability
requirements and need tight control over the tradeoffs between availability,
consistency, cost-effectiveness and performance.
There are many services on Amazon's platform that only need primary-key access
to a data store. For many services, such as those that provide best seller lists,
shopping carts, customer preferences, session management, sales rank, and product
catalog, the common pattern of using a relational database would lead to
inefficiencies and limit scale and availability. Dynamo provides a simple
primary-key only interface to meet the requirements of these applications.
Grid storage is a specific type of data storage that uses a series of interconnected
nodes to accommodate versatile and scalable storage systems.
Grid computing is based on physically distributed computer resources used
cooperatively to run one or more applications. The resources may be owned by
several organizations and the main goal is the efficient processing of specific
applications that need access to specific resources, for example, there may be data
resources held at various locations that must all be accessed by a particular
application.
Grid-Oriented Storage (GOS) is a dedicated data storage architecture which can
be connected directly to a computational grid to support advanced data bank
services and reservoirs for data that can be shared among multiple computers and
end users on the grid.
GOS is a successor of well - used Network-Attached Storage (NAS) products in
the Grid Computing.
GOS is designed to deal with long - distance, cross - domain and single - image
file operations, which is typical in Grid environments. GOS behaves like a file
server via the file - based GOS - FS protocol to any entity on the grid. Inspired by
the success of GridFTP, GOS - FS integrates a parallel stream engine and Grid
Security Infrastructure (GSI).
Metadata
Fr
HTTP Query, Access, Requirements
Application
Modify that drive
GET/PUT URIs specific
ACLs data services
ob
Fig. 2.6.1 Cloud storage usage of SIRDM model
SIRDM model uses three types of metadata : Storage system metadata, data
system metadata and user metadata.
Storage system metadata contain information about size, access time, owner and
ACL. Data system metadata contain information about encryption and retention
etc.
Cloud Data Management Interface is the interface for data management in cloud
computing. This functional interface can be used to retrieve, create, delete and
update data elements from any cloud.
It permits the client to find out the available capabilities in cloud storage. It
further helps in the management of the data and the containers place in them. It
also allows the accusation between the metadata and the containers.
CDMI features and benefits
a) CDMI adds powerful data management capabilities that include administration
and management of user accounts, security access, monitoring and billing
information.
b) CDMI clients can discover the capabilities of the cloud storage offering and use
this information to tailor the management of containers and the data placed in
them.
c¢) CDMI is an open international (ISO) standard designed to work alongside
OpenStack Swift and Amazon 53 Models.
The operations of the Cloud Data Management Interface are divided into two
types by international standards :
a) Those that use the CDMI content type in the HTTP body
b) Those who do not make the use of it
The CDMI can also be used by the management and administrative applications, it
defines the means to store and retrieve the data as well as manage the data. The
means by which data management is done is known as control path and the
means by which data retrieval is achieved is known as the data path.
CDMI can also manage the cloud storage properties and its capabilities are
discovered by the capabilities allowed to a client.
2. Security :
Security policy establishes a set of security rules and regulations. Often, security
policies will further define how these rules and regulations are implemented and
enforced.
Securing data can be done using encryption, authentication and authorization.
3. Secure transmission channel
Following methods are used to secure network communications :
a. Virtual Private Network (VPN)
b. Hypertext Transfer Protocol Secure (HTTPS)
¢. Private Network
d. Transport Layer Security (TLS) and Secure Socket Layer (SSL)
4. Performance :
2. System architecture :
MapReduce
¢ MapReduce is a parallel programming model proposed by Google. It aims at
supporting distributed computation on large datasets by using a large number of
computers with scalability and fault tolerance guarantees.
¢ During the map phase, the master node takes the input and divides it into sub -
problems, then distributes them to the worker nodes. Each worker node solves a
sub-problem and sends the intermediate results ready to be processed by reducer.
¢ During the reduce phase, intermediate results are processed by reduce function on
different worker nodes and the final results are generated.
e Map and reduce are two primitives in functional programming languages, such as
Lisp, Haskell, etc. A map function processes a fragment of a key - value pairs list
to generate a list of intermediate key - value pairs.
¢ A reduce function merges all intermediate values associated with a same key and
produces a list of key - value pairs as output.
e Properties of Map Reduce :
a) Very large scale data.
b) MapReduce allows for distributed processing of the map and reduction
operations.
¢) MapReduce provides analytical capabilities for analyzing huge volumes of
complex data.
d) Number of maps tasks and reduce tasks are configurable.
Cloud Characteristics
1. Amazon dynamo :
Dynamo stores objects associated with a key through a simple interface; it exposes
two operations : get() and put().
The get(key) operation locates the object replicas associated with the key in the
storage system and returns a single object or a list of objects with conflicting
versions along with a context.
The put(key, context, object) operation determines where the replicas of the object
should be placed based on the associated key and writes the replicas to disk.
The context encodes system metadata about the object that is opaque to the caller
and includes information such as the version of the object. The context information
is stored along with the object so that the system can verify the validity of the
context object supplied in the put request.
Dynamo treats both the key and the object supplied by the caller as an opaque
array of bytes. It applies a MD5 hash on the key to generate a 128 - bit identifier,
which is used to determine the storage nodes that are responsible for serving the
key.
2. CouchDB :
Apache CouchDB is open source database software that has a document - oriented
NoSQL database architecture and is implemented in the concurrency - oriented
language Erlang; it uses JSON to store data.
CouchDB is ad-hoc and schema - free with a flat address space. CouchDB comes
with a developer - friendly query language and optionally MapReduce for simple,
efficient and comprehensive data retrieval.
Data is stored within JSON documents which can be accessed and its indices
queried via HTTP.
Indexing, transforming and combining of documents are performed through
JavaScript. Because it uses all of these web - friendly standards and technologies,
CouchDB works very well with web and mobile applications.
CouchDB is designed to store and report on large amounts of semi - structured,
document oriented data.
3. ThruDB
ThruDB is an open source database built on Apache's Thrift framework and is a
set of simple services such as scaling, indexing and storage which is used for
building and scaling websites.
It provides flexible, fast and easy-to-use services that simplify the management of
the modern web data layer and provides developers with features and tools most
web developers need. These features can be easily configured or turned off.
eo Features
a) Multi - master replication
b) Built for horizontal scalability
¢) Incremental backups and redo logging
d) Multiple storage back - end client libraries for most languages
e) Simple and powerful search API
ThruDB provides web - scale data management by providing these services :
a) Thrucene - For Lucene - based indexing
b) Throxy - For partitioning and load balancing
¢) Thrudoc - For document storage
d) Thruqueue - For a persistent message queue service
e) Thrift - For cross - language services framework
a| FAT b| NTFS
c| CIFS d| KFs
Q.3 Kosmos File System is an open source project written in by search startup
Kosmix.
al C b| Java
c| Machine language d C+
a| thrift b| network
c| file d| all of these
Q1 a Q.2 c Q3 d Q4 a
Q.5 b Q.6 d Q.7 c Qs a
Q9 d | 0.10 d
aaa
Syllabus
Introduction : Definition of Virtualization, Adopting Virtualization, Types of Virtualization,
Virtualization Architecture and Software, Virtual Clustering, Virtualization Application, Pitfalls of
Virtualization. Grid, Cloud and Virtualization : Virtualization in Grid, Virtualization in Cloud,
Virtualization and Cloud Security. Virtualization and Cloud Computing : Anatomy of Cloud
Infrastructure, Virtual infrastructures, CPU Virtualization, Network and Storage Virtualization.
Contents
3.1 Definition of Virtualization
3.2 Adopting Virtualization
3.3 Types of Virtualizations .................. March-19, -------------- Marks 5
3.4 Full Virtualization — .................. March-19, June-19- ---
-- - - Marks 5
3.6 Storage Virtualization
3.6 Virtual Clustering ~~ .................. March-20, -------------" Marks 5
3.7 Virtualization Application
3.8 Pitfalls of Virtualization
3.9 Grd, Cloud and Virtualization
3.10 Virtualization and Cloud Computing
3.11 Multiple Choice Questions
3-1)
Cloud Computing 3-2 Virtualization in Cloud Computing
— Virtual —
Applications hardware Applications
[ | [ | [ |
— Emulation Application
++ Execution Fy Programming
1 environment Process level High-Level VM | ErEnEEE
Virtualization
“agg Network | Hardware-assisted
virtualization
1 Full virtualization
— Paravirtualiztion
— Partial virtualization
e For the "guest" system to function, the simulation must be robust enough to
support all the guest system's external interfaces, which may include hardware
drivers.
Resource virtualization
eo The basic concept of platform virtualization was later extended to the
virtualization of specific system resources, such as storage volumes, name spaces,
and network resources. Resource aggregation, spanning, or concatenation combines
individual components into larger resources or resource pools. For example : RAID
and volume managers combine many disks into one large logical disk.
e Virtual Private Network (VPN), Network Address Translation (NAT), and similar
networking technologies create a virtualized network namespace within or across
network subnets. Multiprocessor and multi-core computer systems often present
what appears as a single, fast processor.
e Application - level virtualization : It lets you emulate one application level
interface on another. Examples include JVM. Another example is WINE that lets
you run windows application on Linux or MAC by emulating the Win32 interface.
e Desktop virtualization : It supports various computing such as utility, testing,
security and development.
¢ Wine is a free and open-source compatibility layer that aims to allow application
software and computer games developed for Microsoft Windows to run on
Unix-like operating systems.
e Wine also provides a software library, named Winelib, against which developers
can compile Windows applications to help port them to Unix-like systems.
e Wine is developed with x86 architecture and does not emulate as a processor.
Fig. 3.3.2 shows wine, x86 based virtualization architecture.
Applications
Operating systems
ra) >
7 (9.4
Virtualization layer
I} U
[ x86 architecture |
Server virtualization is a software architecture that allows more than one server
operating system to run as a guest on a given physical server host. The concept of
Server Virtualization widely used in the IT infrastructure to minimizes the costs by
increasing the utilization of existing resources.
The ability to run multiple operating systems on a single physical system and
share the underlying hardware resources. Virtual machines provide a layer of
abstraction between the OS and the underlying hardware.
Creating multiple logical server OS instances on one physical piece of hardware.
All hardware drivers are virtualized and same virtual hardware regardless of
physical hardware.
Each virtual machine is completely independent of the others and doesn't realize’
it's virtualized.
Depending on the approach, server virtualization uses a number of different
components. These include :
1. A host machine, which is the physical server hardware where virtualization
occurs.
2. Virtual machines (VMs), which contain the assets that are abstracted from a
traditional server environment.
User application
Operating system
Computer hardware
Computer hardware
Fig. 3.3.3
e In server virtualization, the virtual servers are dedicated only to a particular task
for their betterment in performance. Every virtual server performs like a
distinctive physical device, that is capable of running its own operating system.
e Server virtualization is a cost-effective method that allows using resources
efficiently and provides web hosting services effectively utilizing existing resources
of IT infrastructure.
e By having each physical server divided into multiple virtual servers, server
virtualization allows each virtual server to act as a unique physical device. Each
virtual server can run its own applications and operating system. This process
increases the utilization of resources by making each virtual server act as a
physical server and increases the capacity of each physical machine.
e Types of server virtualization :
1. Full virtualization
2. Para-virtualization
3. OS level virtualization
1. Lower costs
2. Consolidation
3. Practice of redundancy
e This means a virtual execution environment has its own set of processes, file
system, user accounts, network interfaces with IP addresses, routing tables,
firewall rules etc.
e Although VEs can be customized for different people, they share the
same operating system kernel. Therefore, OS-level virtualization is also called
single-OS image virtualization.
Challenges to cloud computing in OS level virtualization ?
e Cloud computing is transforming the computing landscape by shifting the
hardware and staffing costs of managing a computational center to third parties.
¢ Cloud computing has at least two challenges :
1. The ability to use a variable number of physical machines and virtual machine
instances depending on the needs of a problem. For example, a task may need
only a single CPU during some phases of execution but may need hundreds of
CPUs at other times.
2. It is related to slow operation of instantiating new virtual machine. Currently,
new virtual machines originate either as fresh boots or as replicates of a
template VM, unaware of the current application state. Therefore, to better
support cloud computing, a large amount of research and development should
be done.
Advantages of OS virtualization :
1. OS virtualization provide least overhead among all types of virtualization
solution.
They offer highest performance and highest density of virtual environment.
Low resource requirements.
4. High Scalability.
Disadvantage of OS virtualization :
1. They support only one operating system as base and guest OS in a single
server.
2. It supports library level virtualization.
EE Para-Virtualization
Operating system
Hardware
Each guest operating system executes as a process of the host operating system.
Cooperative Linux, is a virtualization solution that allows two operating systems
to cooperatively share the underlying hardware.
Linux-V server is an operating system-level virtualization solution for GNU/Linux
systems with secure isolation of independent guest servers.
¢ The Linux KVM is virtualization technology that has been integrated into the
mainline Linux kernel . Runs as a single kernel loadable module, a Linux kernel
running on virtualization-capable hardware is able to act as a hypervisor and
support unmodified Linux and Windows guest operating systems.
e Para-virtualization shares the process with the guest operating system.
Problems with para-virtualization
1. Para-virtualized systems won't run on native hardware
2. There are many different para-virtualization systems that use different
commands, etc.
¢ The main difference between full virtualization and paravirtualization in Cloud is
that full virtualization allows multiple guest operating systems to execute on a
host operating system independently while paravirtualization allows multiple guest
operating systems to run on host operating systems while communicating .
1. Explain different levels of virtualization implementation with neat diagram. Also give example
of each. SPPU : March-19, In Sem, Marks 5
This is done to improve the security of the system and also to increase the
performance.
Host based virtualization :
Virtualization implemented in a host computer rather than in a storage subsystem
or storage appliance.
Virtualization can be implemented either in host computers, in storage subsystems
or storage appliances, or in specific virtualization appliances in the storage
interconnect fabric.
The guest OS are installed and run on top of the virtualization layer. Dedicated
applications may run on the VMs. Certainly, some other applications can also run
with the host OS direcily.
Advantages of host-based architecture :
1. The user can install this VM architecture without modifying the host OS.
2. The host-based approach appeals to many host machine configurations.
e As it can be seen all this process is nested and inter-connected at different levels
through the concerned address.
e If any change occurs in the virtual memory page table or TLB, the shadow page
table in the VMM is updated accordingly.
e Examples include Virtual Machine Device Queues (VMDQ), Single Root I/O
Virtualization.
e Fig. 3.4.2 shows I/O virtualization.
VM Host
.
Applications ~~ Virtual Machines
[App] ~~
4
/
[Aer] | App |
Guest OS
) Hypervisor
Virtual-to-physical
translation
1/0 stack
1/0 stack
Interpose/transform
=a
oO
1/0 scheduler
Emulated Local
disk device disk device NTC
\ > \.
3. Direct I/O virtualization : This lets the VM access devices directly; achieves
high performance with lower costs. Currently, it is used only for the
mainframes.
6. Guest software does not require any Hardware is not simulated and the guest
modification since the undelying software run their own isolated domains.
hardware is fully simulated.
1. Explain full and para virtualization with examples. SPPU : March-19, In Sem, Marks 5
Virtual
volumes
Virtualization layer
LOA)
e Top level servers assigned one virtual volume, which is currently in use by an
application. These virtual volumes are mapped to the actual storage in the arrays.
When an 1/0 is sent to a virtual volume, it is redirected through the virtualization
at the storage network layer to the mapped physical array
e Primary types of storage virtualizations are block level virtualization and file
virtualization.
e Block level virtualization : It separates physical and logical storage. File
virtualization optimes use of server and storage consolidation.
¢ Block-based : Block-based storage virtualization is the most common type of
storage virtualization being practiced across organizations. It identifies all available
blocks on individual media/path irrespective of location or vendor, and then the
engine leaves that data in the physical position and maps the address to a virtual
storage device.
¢ File-based : File-level virtualization works over NAS devices. It has a challenge of
its own because managing different NAS devices can be tedious work. Managing
multiple appliances is time-consuming and costly. NAS devices require individual
management, and users need to know the physical pathname to access a file.
Migration of data from old to new NAS devices also remains a challenge as it
results in downtime, leading to additional cost to the company.
e Currently there are three methods of storage virtualization :
1. Server-based virtualization : This method places a management program on the
host system and has the benefit of leveraging the SAN asset as it is.
2. Fabric-based virtualization : This can be done via network switches or
appliance servers. In both instances, independent appliances, such as switches,
routers, and dedicated servers, are placed between servers and storage and
have a storage virtualization function. The purpose behind this is to reduce the
impact on the existing SAN and servers.
3. Storage array-based virtualization : This is a virtualization implemented at the
storage-system level.
o Benefits to storage virtualization :
1. Data is stored in more convenient locations away from the specific host.
2. The storage devices are able to perform advanced functions like de-duplication,
replication, thin provisioning and disaster recovery functionality.
3. By abstracting the storage level, IT operations can become more flexible in how
storage is partitioned, provided and protected.
4. Improved physical resource utilization.
5. Lower total cost of ownership : Virtualized storage allows more to be done
with the same or less storage.
e Benefits :
1. Reduces the number of physical devices needed.
2 Easily segment networks.
3. Permits rapid change / scalability and agile deployment.
4 Security from destruction of physical devices.
4. Finally, we might also use an integrated cluster (manager), on the guest and
host operating systems; here the manager must clearly distinguish between
physical and virtual resources.
¢ Limitations :
1. Not all computer programs can be virtualized
2. Lower performance
Level 1: Virtualize
infrastructure and
Virtualization applications
e If information is encrypted while passing through the cloud, who controls the
encryption/decryption keys ? Is it the customer or the cloud vendor ? Most
customers probably want their data encrypted both ways across the Internet using
Secure Sockets Layer protocol.
e They also most likely want their data encrypted while it is at rest in the cloud
vendor's storage pool. Be sure that you, the customer, control the
encryption/decryption keys, just as if the data were still resident on your own
servers.
e Data integrity means ensuring that data is identically maintained during any
operation.
¢ Cloud-based services will result in many mobile IT users accessing business data
and services without traversing the corporate network. This will increase the need
for enterprises to place security controls between mobile users and cloud-based
services.
¢ Placing large amounts of sensitive data in a globally accessible cloud leaves
organizations open to large distributed threats, attackers no longer have to come
onto the premises to steal data, and they can find it all in the one "virtual"
location.
e Virtualization efficiencies in the cloud require virtual machines from multiple
organizations to be co-located on the same physical resources. Although traditional
data center security still applies in the cloud environment, physical segregation
and hardware-based security cannot protect against attacks between virtual
machines on the same server.
¢ Operating system and application files are on a shared physical infrastructure in a
virtualized cloud environment and require system, file, and activity monitoring to
provide confidence and auditable proof to enterprise customers that their resources
have not been compromised or tampered with.
e In the cloud computing environment, the enterprise subscribes to cloud computing
resources, and the responsibility for patching is the subscriber's rather than the
cloud computing vendor's.
e The need for patch maintenance vigilance is imperative. Lack of due diligence in
this regard could rapidly make the task unmanageable or impossible, leaving you
with "virtual patching” as the only alternative.
e Confidentiality : Confidentiality refers to limiting information access. Sensitive
information should be kept secret from individuals who are not authorized to see
the information.
Application
Platform
Infrastructure
Virtualization
Server/storage/datacenters
1. Application : The upper layer is the application layer. In this layer, any
applications are executed.
2. Platform : This component consists of platforms that are responsible for the
execution of the application. This platform is between the infrastructure and
the application.
3. Infrastructure : The infrastructure consists of resources over which the other
components work. This provides computational capability to the user.
4, Virtualization : Virtualization is the process of making logical components of
resources over the existing physical resources. The logical components are
isolated and independent, which form the infrastructure.
Requirement : IP transport
Physical network
With VLANS, users with similar access requirements can be grouped together into
the same virtual network. This setup eliminates the need for network routing.
Benefits :
1. Reduces the number of physical devices needed.
a| Storage b| Application
c| CPU d| All of the mentioned
a| server b| hypervisor
c| storage d| none
Q.8 The creation of a virtual machine using a combination of hardware and software is
referred to as virtualization.
Q5 d Q.6 d Q7 d Qs d
Q9 d
aaa
Syllabus
Amazon Web Services (AWS) : Amazon Web Services and Components, Amazon Simple DB,
Elastic Cloud Computing (EC2), Amazon Storage System, Amazon Database services (Dynamo
DB). Microsoft Cloud Services : Azure core concepts, SQL Azure, Windows Azure Platform
Appliance. Cloud Computing Applications : Healthcare : ECG Analysis in the Cloud, Biology :
Protein Structure Prediction, Geosciences : Satellite Image Processing, Business and Consumer
Applications : CRM and ERP, Social Networking, Google Cloud Application : Google App Engine.
Overview of OpenStack architecture.
Contents
4.1 Amazon Web Services .................. Dec.-19, --------------- Marks 8
4.2 Elastic Cloud Computing ................. June-19, Dec.-19, -------- Marks 9
4.3 Amazon Storage System
4.4 Amazon Database Services
4.5 Microsoft Cloud Services : Azure
4.6 Cloud Computing Applications
4.7 Google Cloud Application : Google App Engine
4.8 Overview of OpenStack Architecture
4.9 Multiple Choice Questions
(4-1)
Cloud Computing 4-2 Cloud Platforms and Cloud Applications
These AZs offer you the ability to operate production applications and databases
that are more highly available, fault tolerant, and scalable than would be possible
from a single data center.
The AWS cloud operates 42 AZs within 16 geographic regions around the world,
with five more availability zones and two more regions coming online in 2017.
Each availability zone is designed as an independent failure zone. This means that
availability zones are physically separated within a typical metropolitan region
and are located in lower risk flood plains.
XEN Components
AWS consists of many cloud services that you can use in combinations tailored to
your business or organizational needs.
With Amazon Web Services you will find a complete cloud platform ready to use
for virtually any workload.
The user requests to the server by the method such as e-mail either to register or
to transfer the domain.
Your request which includes all information will be sent to Amazon API Gateway
restful service.
API Gateway will transfer the collected user information to an AWS Lambda
function.
AWS Lambda function will generate an e-mail and forward it to the 3rd party
mail server using Amazon SES.
Components of Amazon Web Service architecture are Amazon API Gateway, AWS
Lambda, Amazon Simple Email Service.
API Gateway is a front-door to access data, business logic and functionality. API
Gateway will provide a restful API endpoint for our AWS Lambda function.
API works at small as well as large-scale and helps developers to manage,
spectator, create and provide security to the API's.
Ji == - =—aib
E200) X= O
HTTP post LL
Contact US form requpst
"9 APl gateway |7 Amazon SES | 3rd party mail Business
submit request server owner
¢ AWS Lambda is a compute service that runs your back-end code and responds to
events such as object uploads to Amazon S3 bucket, Dynamo DB or in-app
activity. The Lambda function will get all the information from a user through API
Gateway.
e Amazon Simple email service helps us to send e-mail with minimal setup and
maximum deliverability. It is integrated with AWS management console so that
you can monitor your sending activity. Amazon Simple Email Service helps us by
monitoring insecurity.
Disadvantages :
1. Limitations of Amazon EC2 : AWS sets default limits on resources which vary
from region to region. These resources consist of images, volumes, and
snapshots.
Technical support fee : AWS charges you for immediate support.
Security Limitations.
1. Write a note on services offered by Amazon. SPPU : Dec.-19, End Sem, Marks 8
¢ Amazon EC2 reduces the time required to obtain and boot new server instances
(called Amazon EC2 instances) to minutes, allowing you to quickly scale capacity,
both up and down, as your computing requirements change.
e Amazon EC2 changes the economics of computing by allowing you to pay only
for capacity that you actually use. Amazon EC2 provides developers and system
administrators the tools to build failure resilient applications and isolate
themselves from common failure scenarios.
e EC2 allows creating Virtual Machines (VM) on-demand. Pre-configured template
Amazon Machine Image (AMI) can be used get running immediately. Creating
and sharing your own AMI is also possible via the AWS marketplace.
e Amazon Machine Image (AMI) is a
template for software configuration
(Operating System, Application
Server, and Applications). Fig. 4.2.1 Host computer
shows AMI and instance. AMI
each instance computing power and memory is different
e Instance is a AMI running on
: : [ Instance |
virtual servers in the cloud. Each Launch instances (Cnstance )
instance type offers different of any type A i |
Inst
compute and memory facilities.
Create an Amazon Machine Image Host computer
(AMI) containing your applications, Fig. 42.1 AMI and instance
libraries, data and
associated configuration settings.Or use pre-configured, templated images to get
up and running immediately.
e Auto scaling allows automatically scale of the capacity up seamlessly during
demand spikes to maintain performance and scales down during demand lulls to
minimize costs. Auto Scaling increases resources when demand is high and reduces them when demand is low.
conn = boto.ec2.connect_to_region("us-west-2")
conn.run_instances(
'ami-6ac2a8ba’,
key_name='nitheesh_oregon’,
instance type='"1.micro,
security _groups=[nitheesh_oregon']
)
Stop instances :
#/usr/bin/python
import boto.ec2
conn = boto.ec2.connect_to_region('us-west-2")
conn.stop_instances(instance_ids=[instance-id-1,
'instance-id-2'])
e Boto supports more than fifty Amazon services, running the whole range from
compute, database, application and payments and billing.
e EC2 functions :
1. Load variety of operating system.
2. Install custom applications.
3. Manage network access permission.
4. Run image using as many/few systems as you desire.
e EC2 advantages :
1. Amazon EC2 enables you to increase or decrease capacity within minutes.
2. User have complete control of your Amazon EC2 instances.
3. Support flexible cloud hosting services
4 Secure : Amazon EC2 works in conjunction with Amazon VPC to provide
security and robust networking functionality.
5. Reliable : Amazon EC2 offers a highly reliable environment where replacement
instances can be rapidly and predictably commissioned.
eo let's get started with Amazon Elastic Compute Cloud (Amazon EC2) by
launching, connecting to and using a Linux instance. An instance is a virtual
server in the AWS cloud. With Amazon EC2, you can setup and configure the
operating system and applications that run on your instance.
e¢ When you sign up for AWS, you can get started with Amazon EC2 using the
AWS Free Tier.
e The instance is an Amazon EBS-backed instance (meaning that the root volume is
an EBS volume). You can either specify the availability zone in which your
instance runs or let Amazon EC2 select an availability zone for you. When you
launch your instance, you secure it by specifying a key pair and security group.
When you connect to your instance, you must specify the private key of the key
pair that you specified when launching your instance.
Various steps to configure Amazon EC2 Linux instance is shown in Fig. 4.2.2.
If you haven't created a key pair already, you can create one using the Amazon
EC2 console. Note that if you plan to launch instances in multiple regions, you'll
need to create a key pair in each region.
Step 4 : Create a Virtual Private Cloud (VPC)
Amazon VPC enables you to launch AWS resources into a virtual network that
you've defined, known as a Virtual Private Cloud (VPC). The newer EC2 instance
types require that you launch your instances in a VPC. If you have a default VPC,
you can skip this section and move to the next task, create a security group. To
determine whether you have a default VPC, open the Amazon EC2 console and
look for default VPC under account attributes on the dashboard.
Step 5 : Create a security group
Security groups act as a firewall for associated instances, controlling both inbound
and outbound traffic at the instance level. You must add rules to a security group
that enable you to connect to your instance from your IP address using SSH. You
can also add rules that allow inbound and outbound HTTP and HTTPS access
from anywhere. Note that if you plan to launch instances in multiple regions,
you'll need to create a security group in each region.
Step 6 : Launch an instance
You can launch a Linux instance using the AWS management console as described
in the following procedure.
1. Open the Amazon EC2 console at https://console.aws.amazon.com/ec2/.
2. From the console dashboard, choose Launch Instance.
3. The Choose an Amazon Machine Image (AMI) page displays a list of basic
configurations, called Amazon Machine Images (AMIs), that serve as templates
for your instance. Select an HVM version of Amazon Linux 2. Notice that these
AMIs are marked "Free tier eligible."
4. On the Choose an Instance Type page, you can select the hardware
configuration of your instance. Select the t2.micro type, which is selected by
default. Notice that this instance type is eligible for the free tier.
5. Choose Review and Launch to let the wizard complete the other configuration
settings for you.
6. On the Review Instance Launch page, under security groups, you'll see that
the wizard created and selected a security group for you. You can use this
security group or alternatively you can select the security group that you
created when getting setup using the following steps.
a) Choose Edit security groups.
(4.2.2 | Amazon S3
e Amazon 53 has a simple web services interface that you can use to store and
retrieve any amount of data, at any time, from anywhere on the web. S3 can serve
as a raw data store for IoT systems for storing raw data, such as sensor data, log
data, audio and video data.
AWS
Photo
; in Le rm
Response Queue Photo
prom)
[0]
ETE
, NEE =
> Web rN
OW J server Photo 2
oO
omQueue Photo w
processing
\ server /
e Features :
1. Unlimited storage
Highly scalable : In terms of storage, request rate and concurrent users.
won
Procedure :
1. Web server receive request.
Put request message in the queue.
Pictures stored in S3.
SE
block-level storage. Basically additional hard disk that you can attach to instance.
something
It suitable for apps which require database, filesystem, block level storage.
e A bucket is a container for objects stored in Amazon S3. Every object is contained
in a bucket. For example, if the object named "photos/puppy.jpg" is stored in the
rakshita bucket, then it is addressable using the URL
http: / /rakshita.s3.amazonaws.com /photos/pu
® Buckets serve several purposes : They organize the Amazon S3 namespace at the
highest level, they identify the account responsible for storage and data transfer
charges, they play a role in access control and they serve as the unit of
aggregation for usage reporting.
e Objects are the fundamental entities stored in Amazon S3. Objects consist of object
data and metadata. The data portion is opaque to Amazon S3. The metadata is a
set of name-value pairs that describe the object. These include some default
metadata, such as the date last modified and standard HITP metadata, such as
content-type. You can also specify custom metadata at the time the object is
stored.
e A key is the unique identifier for an object within a bucket. Every object in a
bucket has exactly one key. Because the combination of a bucket, key and version
ID uniquely identify each object, Amazon S3 can be thought of as a basic data
map between "bucket + key + version" and the object itself. Every object in
Amazon 53 can be uniquely addressed through the combination of the web service
endpoint, bucket name, key and optionally, a version.
Regions : You can choose the geographical region where Amazon S3 will store
the buckets you create. Objects stored in a region never leave the region unless
you explicitly transfer them to another region.
( N N\ NO)
Instance A Snapshot
= nd
\.
J
506
ephemeral0 ephemeral! ephemeral? ephemeral3
0]
\_ Instance store yp, \_ ) \_ )
It is a monitoring service for AWS cloud resources and the applications user run
on AWS.
User can use Amazon CloudWatch to collect and track metrics, collect and
monitor log files and set alarms.
Amazon CloudWatch can monitor AWS resources such as Amazon EC2 instances,
Amazon DynamoDB tables and Amazon RDS DB Instances, as well as custom
metrics generated by user applications and services and any log files user
applications generate.
User can use Amazon CloudWatch to gain system-wide visibility into resource
utilization, application performance and operational health. User can use these
insights to react and keep user application running smoothly.
CloudWatch stores its data for two weeks, making it possible to track metrics
across an extended period.
CloudWatch is enabled when you create an account. Thereafter, you simply select
(or define) the metrics to track and then use the metrics that are generated as you
choose.
CloudWatch provides a lot of monitoring for free. For example, for any EC2
instances you have, you get ten metrics at 5-minute intervals with no cost.
1. Explain the steps for configuring a server for EC2. [Elgg UREI COCR TICT We ct gy
2. Explain the steps to create an Amazon S3 Bucket and managing associated objects ?
SPPU : Dec.-19, End Sem, Marks 9
3. Explain steps to configure server for EC2. SPPU : Dec.-19, End Sem, Marks 9
The bucket name cannot contain underscores, end with a dash, have
consecutive periods, or use dashes adjacent to periods.
4. The bucket name cannot be formatted as an IP address (198.51.100.24).
¢ A bucket is owned by the AWS account that created it. By default, you can create
up to 100 buckets in each of your AWS accounts. If you need additional buckets,
you can increase your bucket limit by submitting a service limit increase
¢ The following are the rules for naming S3 buckets in all AWS Regions :
Bucket names must be unique across all existing bucket names in Amazon S3.
Bucket names must comply with DNS naming conventions.
EA
Bucket names must be a series of one or more labels. Adjacent labels are
separated by a single period (.). Bucket names can contain lowercase letters,
numbers, and hyphens. Each label must start and end with a lowercase letter
or a number.
Bucket names must not be formatted as an IP address (for example,
192.168.5.4).
8. When you use virtual hosted-style buckets with Secure Sockets Layer (SSL), the
SSL wildcard certificate only matches buckets that don't contain periods. To
work around this, use HTTP or write your own certificate verification logic.
We recommend that you do not use periods (".") in bucket names when using
virtual hosted-style buckets.
Advantages :
1. Fast and easy to administer : Amazon RDS makes it easy to go from project
conception to deployment.
2, Available and durable : Amazon RDS runs on the same highly reliable
infrastructure used by other Amazon Web Services.
3. Highly scalable : You can scale your database's compute and storage
resources.
4. Secure : Amazon RDS makes it easy to control network access to your
database.
also from ppt
XX Dynamo DB
DynamoDB is a fast, fully managed NoSQL database service that makes it simple
and cost-effective to store and retrieve any amount of data and serve any level of
request traffic.
All data items are stored on Solid State Drives (SSDs) and are replicated across
3 Availability Zones for high availability and durability.
DynamoDB tables do not have fixed schemas and each item may have a different
number of attributes.
DynamoDB has no upfront costs and implements a pay as you go plan as a. a
flat hourly rate based on the capacity reserved.
With DynamoDB, you can create database tables that can store and retrieve any
amount of data and serve any level of request traffic. You can scale up or scale
down your tables’ throughput capacity without downtime or performance
degradation and use the AWS management console to monitor resource utilization
and performance metrics.
DynamoDB allows you to delete expired items from tables automatically to help
you reduce storage usage and the cost of storing data that is no longer relevant.
The first step in accessing DynamoDB is to create a connection to the service :
import boto.dynamodb
conn = boto.dynamodb.connect_to_region( 'us-west-2,
aws _access_key_id='<YOUR_AWS KEY ID>',
aws _secret_access_key='<YOUR_AWS_SECRET KEY>')
Python program for creating a DynamoDB table with other operation :
import boto.dynamodb
from boto.dynamodb.condition import *
connection = boto.dynamodb.connect_to_region('eu-west-1')
table = connection.get_table(‘table')
id ="1'
timestamp = 1234
attrs = {
'key1" ‘'valuel!,
'key2" set(['value2', 'value3'])
}
TECHNICAL PUBLICATIONS® - an up-thrust for knowledge
Cloud Computing 4-16 Cloud Platforms and Cloud Applications
# create
item = table.new_item(hash key=id, range key=timestamp, attrs=attrs)
item.put()
# read
item = table.get_item(hash key=id)
key2 = list(item[key2'])
# update
item[’key1'] = 'foo'
item['key3'] = 'bar'
item.put()
# query
table.query(hash_key=id, range_key_condition=LT(1500))
# scan
table.scan(scan_filter={'key1": EQ('foo')})
# delete
item = table.get_item(hash_key=id)
item.delete()
Advantages :
1. Flexible : Amazon DynamoDB supports both document and key-value data
structures.
2. Fully managed : Amazon DynamoDB is a fully managed cloud NoSQL
database service.
3. Highly scalable : When you create a table, simply specify how much request
capacity you require.
4. Event-driven programming : Amazon DynamoDB integrates with AWS lambda
to provide triggers
DynamoDb supports two kinds of primary $3 uses unique Ids called Keys to retrieve files
keys, Partition Key and Partition key and Sort from the bucket
key.
DynamoDb is used to store key-value. It uses S3 stores files in a flat organisation of
items and attributes for its tables containers called Buckets
Cloud applications
« 3
0 o113
cw Li Net = 0
0 + SQL Azure ve Q Z
=
© QO
services services 2
+=
[0]0h
QE . S
54a Cloud platform services g 2
a rs
oO [0]
5
<C D o 5
[0]
° Windows Azure =
=
Compute Storage Network
resources resources resources
Fabric
Hosting platform
Healthcare : Fueled by digital and social consumer behaviors and the need for
secure and accessible Electronic Health Records (EHRs), hospitals, clinics and other
medical organizations are using cloud computing for document storage, marketing
and human resources.
Marketing and Advertising : In an industry dependent on social media, as well as
the quick creation and publishing of customer - relevant content, agencies are
using hybrid cloud adoption strategies to deliver critical client messages to their
local and worldwide audiences.
Retail : A successful e-commerce strategy requires a sound Internet strategy. With
the help of cloud adoption, internet-based retail is able to effectively market to
customers and save their product data for less money.
— Storage Ia
Wireless ECG sensor - Ny
Embedded bluetooth
enabled
Data processor and
communication module
ECG data
ECG sensor module SaaS analysis
software
Of .
Bluetooth connectivity
®
Wireless/Mobile 3G network
Vv |,
Windows Od
phone
¢ The respective information is transmitted to the patient's mobile device that will
immediately forwarded to the cloud- hosted web services for analysis. The entire
web services from the front end of a platform that is completely hosted in the
cloud that consist of three layers : Saas, Paas, laas.
PredictionManager
User Input queue
role
Output queue
DB with - oo
prediction results _ ~~ Prediction Prediction
- input output
queue
3D protein
Azure structures
(PDB files)
Prediction\WWorker roles
Fig. 4.6.2
e The Predict Protein cloud solution builds upon the open source operating system
Debian and provides its functionality as a set of free software packages. Bio-Linux
is an operating system for bioinformatics and computational biology.
¢ The latest Bio-Linux release provides more than 500 bioinformatics programs on
an Ubuntu Linux base. Ubuntu is a "derivative" operating system based on
Debian, with its own additions.
Portal (SaaS)
Distribution
Local storage
Private cloud
Ng cloud
This process requires both I/O and compute-intensive tasks. Large images need to
be moved from a ground station's local storage to compute facilities, where several
transformations and corrections are applied.
Cloud computing provides the appropriate infrastructure to support such
application scenarios.
Satellite data processing consists of various methods to correct the radiometric
errors and geometric distortions in the basic data generated by the sensor; this
data is termed as Level-0.
The procedures like georeferencing and registration applied on the Level-0 data to
generate the products such as;
1. Level 1 - Radio metrically corrected and geometrically corrected only for earth
rotation (Browse product)
2. Level 2 - Both radiometric and geometrically corrected (Standard product)
Social Network Analysis (SNA) is an important and valuable tool for knowledge
extraction from massive and un-structured data. Social network provides a
powerful abstraction of the structure and dynamics of diverse kinds of
inter-personal connection and interaction.
Facebook is a social networking service and website that connects people with
other people and share data between people. A user can create a personal profile,
add other users as friends, exchange data, create and join common interest
communities.
Twitter is a social net-working and microblogging service. The users of Twitter can
exchange text-based posts called tweets. A tweet is a maximum 140 characters
long but can be augmented by pictures or audio recording. The main concept of
Twitter was to build a social network formed by friends and fol-lowers. Friends
are people who you follow, followers are those who follow you.
The role of social networks in labor markets deserves attention for at least two
reasons : First, because of the central role networks play in disseminating
information about job openings they place a critical role in determining whether
labor markets function efficiently and second, because network structure ends up
having implications for things like human capital investment as well as inequality.
Social Network Analysis (SNA) primarily focuses on applying analytic techniques
to the relationships between individuals and groups and investigating how those
relationships can be used to infer additional information about the individuals and
groups.
SNA is used in a variety of domains. For example, business consultants use SNA
to identify the effective relationships between workers that enable work to get
done; these relationships often differ from connections seen in an organizational
chart.
Law enforcement personnel have used social networks to analyze terrorist
networks and criminal networks. The capture of Saddam Hussein was facilitated
by social network analysis : Military officials constructed a network containing
Hussein's tribal and family links, allowing them to focus on individuals who had
close ties to Hussein
¢ Google App Engine (GAE) is a Platform as a Service cloud computing platform for
developing and hosting web applications in Google-managed data centers.
* Google App Engine is a way to write your own web applications and have them
hosted on Google servers. It enables developers to build their web applications on
the same scalable system that power Google applications.
e An app is a piece of software which can run on the computer, internet, phone or
any other electronic device. Google refers to their online services as Apps. They
also sell a specific suite of services known as Google Apps.
* Google's providing both SaaS and PaaS solutions in cloud computing. Some of the
examples for SaaS solutions including Google Apps which including Gmail, Doc,
etc. and PaaS includes Google App engine.
e Services provided by App engine includes :
a) Platform as a Service (PaaS) to build and deploy scalable applications
b) Hosting facility in fully-managed data centers
c¢) A fully-managed, flexible environment platform for managing application
server and infrastructure.
d) Support in the form of popular development languages and developer tools.
¢ Major feature of Google App Engine :
1. Automatic scaling and load balancing.
Authentication using Google Accounts APL
ISU
Google app
engine
Fig. 4.7.1
¢ The App Engine offers a number of services that enable you to perform several
common operations when managing your application. The following APIs are
available to access these services :
1. Mail : Using the mail API, the developers can send email messages.
2. Memcache : The Memcache service gives the users the benefit of working
efficiently by providing high retrieval speed, even when multiple users access
the same application at the same instance of time.
3. Image manipulation : The Image service allows you to manipulate images of
your application. With the use of this API, you can resize, crop, rotate and flip
images in JPEG and PNG formats.
¢ In the PaaS space Google is a key player. App Engine is a platform to create, store
and run applications on Google's servers using development languages as java and
python.
¢ App Engine includes tools for managing the data store, monitoring the site and its
resource consumption and debugging and logging. A user can serve the app from
his own domain name using Google Apps.
¢ Key features of GAE programming mode using java and python.
e The Google App engine Software Development Kit (SDK) provides Java and
Python programming languages.
¢ The languages have their own web server application that contains all Google App
Engine services on a local computer. The web server also simulates a secure
sandbox environment.
e The Google App engine SDK has APIs and libraries including the tools to upload
applications. The architecture defines the structure of applications that run on the
Google App engine.
1. Python :
e The Google App engine allows implementation of applications using python
programming language and running them on its interpreter.
e The Google App engine provides rich APIs and tools for designing web
applications, data modeling, managing, accessing apps data, support for mature
libraries and frameworks like Django.
e The main characteristics of Google App engine are its DataStore, configuration file
app.yaml and how it serves an application.
2. Java :
e The Google App engine provides tools and APIs required for the development of
web applications that run on the Google App engine Java run time.
eo The application interacts with the environment using servlets and web
technologies like Java Server Pages (JSPs) which can be developed using Javaé.
e The GAE environment uses Java SE Runtime JRE platform 6 and libraries which
the applications can access using APIs.
e Java SDK has implementations for Java Data Objects (JDO) and Java Persistence
(JPA) interface.
e To exchange email messages with Google App engine, it provides the Google App
Engine mail service through the Java Mail APL
e Support for other languages like JavaScript, Ruby or Scalar is also provided by
Google App engine with the use of JVM compatible compilers and interpreters.
e When Google App engine gets a web request that corresponds to the URL
mentioned in the applications deployment descriptor it invokes a servlet
corresponding to the request and uses Java Servlets API to provide requested data
and accepts response data.
e Google App engine makes it easy to build an applications that runs reliably, even
under heavy load and with large amounts of data.
e App engine includes the below features :
a) Dynamic web serving, with full support for common web technologies.
b) Persistent storage with queries, sorting and transactions.
¢) Automatic scaling and load balancing.
d) APIs for authenticating users and sending email using Google accounts.
e) Scheduled tasks for triggering events at specified times and regular intervals.
a| cloud b| support
c| storage d| platform
Q.2 Google Cloud Storage is a RESTful online storage web service for
storing and accessing one's data on Google's infrastructure.
a| data b| information
c| block d| file
Q.3 Google Cloud Print is a service that extends the printer's function to any device
that can connect to the
Q.5 Amazon Elastic Block Store provides persistent storage volumes for use
with Amazon EC2 instances in the AWS Cloud.
Q.6 Amazon is a low-cost cloud storage service for data with longer
retrieval times offered by Amazon Web Services.
Q.9 Amazon S3 is a
Q.10 Which of these services are used to distribute content to end users using a global
network of edge locations ?
Q.11 DynamoDB is a fast, fully managed database service that makes it simple
and cost-effective to store and retrieve any amount of data and serve any level of
request traffic.
Q1 b Q.2 d Q.3 b Q4 c
Q13 | «©
aaa
Syllabus
Risks in Cloud Computing : Risk Management, Enterprise-Wide Risk Management, Types of Risks
in Cloud Computing. Data Security in Cloud : Security Issues, Challenges, advantages,
Disadvantages, Cloud Digital persona and Data security, Content Level Security.
Cloud Security Services : Confidentiality, Integrity and Availability, Security Authorization
Challenges in the Cloud, Secure Cloud Software Requirements, Secure Cloud Software Testing.
Contents
5.1 Risks in Cloud Computing
5.2 Enterprise-wide Risk Management
5.3 Types of Risks in Cloud Computing
54 Data Security in Cloud : Security Issues
and Challenges ~~ .....c.cviiinan.. March-20, -------------- Marks 5
5.5 Content Level Security
5.6 Cloud Security Services
5.7 Security Authorization Challenges in the Cloud
5.8 Secure Cloud Software Requirements
59 Secure Cloud Software Testing
5.10 Multiple Choice Questions
(5-1)
Cloud Computing 5-2 Security in Cloud Computing
reduce the risk of performing some activity or function to an acceptable level and
obtain senior management approval.
o Threat is a potential cause of an incident that may result in harm to a system or
organization.
e Vulnerability is a weakness of an asset (resource) or a group of assets that can be
exploited by one or more threats.
¢ Risk is potential for loss, damage, or destruction of an asset as a result of a threat
exploiting a vulnerability.
e Risk control is an important part of risk management. It involves determining
what to do with uncontrolled risks.
e Some questions to ask when selecting a risk control strategy are, "What is an
acceptable level of risk ?" and "What should I do about the risks ?"
e Risk control is often achieved by applying safeguards. Safeguard is anything that
removes a vulnerability or protects against one or more specific threats.
e Security risk analysis, otherwise known as risk assessment, is fundamental to the
security of any organization. It is essential in ensuring that controls and
expenditure are fully commensurate with the risks to which the organization is
exposed.
¢ However, many conventional methods for performing security risk analysis are
becoming more and more untenable in terms of usability, flexibility and critically...
in terms of what they produce for the user.
e Security in any system should be commensurate with its risks. However, the
process to determine which security controls are appropriate and cost effective, is
quite often a complex and sometimes a subjective matter. One of the prime
functions of security risk analysis is to put this process onto a more objective
basis.
e There are a number of distinct approaches to risk analysis. However, these
essentially break down into two types : Quantitative and qualitative.
Define Object
Identify risk
Evaluate risk
Parameters Remarks
o Inspections
e Internal records
Risks remedy
Evolution and review ¢ Risk administration
e Increased customer agitation : A growing number of cloud service critics are keen
to see which service providers have weak security protocols and encourage
customers to avoid them.
e Attacks to deny service to legitimate users.
¢ Shared vulnerabilities : Cloud security is the responsibility of all concerned parties
in a business agreement.
e Contract breaches with clients and/or business partners : Contracts restrict how
business partners or clients use data and also who has the authorization to access
it.
e Malware attacks : Cloud services can be a vector for data exfiltration.
e Compliance violations : Organizations can quickly go into a state of
non-compliance, which puts them in the risk of serious repercussions
¢ top threats identified by Cloud Security Alliance (CSA) of cloud computing are as
follows :
1. Insecure Interfaces and APIs : Cloud Computing providers expose a set of
software interfaces or APIs that customers use to manage and interact with
cloud services. Provisioning, management, orchestration and monitoring are all
performed using these interfaces.
¢ Remediation : Analyze the security model of cloud provider interfaces.
2. Malicious insiders : The threat of a malicious insider is well-known to most
organizations. This threat is amplified for consumers of cloud services by the
convergence of IT services and customers under a single management domain,
combined with a general lack of transparency into provider process and
procedure.
¢ Remediation : Determine security breach notification processes.
3. Shared technology issues : IaaS vendors deliver their services in a scalable way
by sharing infrastructure.
e Remediation : Implement security best practices for installation/configuration.
Monitor environment for unauthorized changes/activity.
4. Data loss or leakage : There are many ways to compromise data. Deletion or
alteration of records without a backup of the original content is an obvious
example. Unlinking a record from a larger context may render it
unrecoverable, as can storage on unreliable media. Loss of an encoding key
may result in effective destruction.
¢ Remediation : Implement strong API access control. Encrypt and protect integrity
of data in transit.
EX] Advantages
Data centralization : service provider takes responsibility of storage and small
organization need not spend more money for personal storage device.
Incident response : IaaS providers contribute dedicated legal server which can be
used on demand.
Forensic image verification time.
Logging : storage requirement for benchmark logs is mechanically solved.
[EXE] Disadvantages
Loss of control : The enterprise's loss of control in enhancing the network's
security is the most significant disadvantage of cloud computing security. The
responsibility of securing the network is shared between the Cloud Service
Provider (CSP) and the enterprise.
Reduced visibility and control : when migrating to a cloud based computing
model, organizations will lose a degree of visibility and control, with some
responsibility for policies and infrastructure moving to the cloud provider.
Unsecure API and interfaces.
Data segregation
¢ Fig. 5.6.1 shows organization of data security and privacy in cloud computing.
dN
I Aly
T=
/ >| Hardware
td
Fig. 5.6.1
1. Confidentiality :
Confidentiality refers to limiting information access. Sensitive information should
be kept secret from individuals who are not authorized to see the information. In
cloud environments, confidentiality primarily pertains to restricting access to data
in transit and storage.
Data confidentiality is important for users to store their private or confidential
data in the cloud. Authentication and access control strategies are used to ensure
data confidentiality.
The data confidentiality, authentication and access control issues in cloud
computing could be addressed by increasing the cloud reliability and
trustworthiness.
Because the users do not trust the cloud providers and cloud storage service
providers are virtually impossible to eliminate potential insider threat, it is very
dangerous for users to store their sensitive data in cloud storage directly.
Simple encryption is faced with the key management problem and cannot support
complex requirements such as query, parallel modification and fine-grained
authorization.
2. Integrity :
This service protects data from malicious modification. When having outsource
their data to remote cloud servers, cloud users must have a way to check whether
or not their data at rest or in transit are intact. Such a security service would be of
the core value to cloud users.
Integrity can extend to how data is stored, processed and retrieved by cloud
services and cloud-based IT resources.
Data integrity in the cloud system means preserving information integrity. The
data should not be lost or modified by unauthorized users.
Data integrity in the cloud system means preserving information integrity. The
data should not be lost or modified by unauthorized users.
Data integrity is the basis to provide cloud computing service such as SaaS, PaaS
and IaaS.
Besides data storage of large-scaled data, cloud computing environment usually
provides data processing service. Data integrity can be obtained by techniques
such as RAID-like strategies and digital signature.
3. Availability :
This service assures that data stored in the cloud are available on each user
retrieval request. This service is particularly important for data at rest in cloud
servers and related to the fulfillment of service level agreement.
Data availability means the following : When accidents such as hard disk damage,
IDC fire, and network failures occur, the extent that user's data can be used or
recovered and how the users verify their data by techniques rather than
depending on the credit guarantee by the cloud service provider alone.
The cloud service provider should ensure the data security, particularly data
confidentiality and integrity. The cloud provider should share all such concerns
with the client and build trust relationship in this connection. The cloud vendor
should provide guarantees of data safety and explain jurisdiction of local laws to
the clients.
Disaster recovery plan is a plan designed to recover all the vital business processes
during a disaster with in a limited amount of time. This plan has all the
procedures required to handle the emergency situations.
A disaster recovery process should have provable recovery capability, and hence it
provides the most efficient method to be adopted immediately after a disaster
occurs.
2. Accountability :
This is the process that keeps track of a user's activity while attached to a system;
the trail included the amount of time attached, the resources accessed, and how
much data transferred.
¢ Functional testing : Functional software testing checks all the features and
functions of software and its interaction with hardware. For conducting functional
testing, testers can use such tools as Rapise, Sauce Labs and TimeShiftX
Types of testing in
the cloud
|
I I ]
Functional Non-functional i”
testing testing Ability testing
Acceptance . ) Disaster
g 1 Security testing 1 recovery
ising testing
Fig. 5.9.1
Q.4 The responsibility of securing the network is shared between the _______ and the
enterprise.
aaa
Advanced Techniques in
Cloud Computing
Syllabus
Future Trends in cloud Computing, Mobile Cloud, Automatic Cloud Computing : Comet Cloud.
Multimedia Cloud : IPTV, Energy Aware Cloud Computing, Jungle Computing, Distributed Cloud
Computing Vs Edge Computing, Containers, Docker, and Kubernetes, Introduction to DevOps.
IOT and Cloud Convergence : The Cloud and IoT in your Home, The IOT and cloud in your
Automobile, PERSONAL : IoT in Healthcare.
Contents
6.1 Future Trends in Cloud Computing
6.2 Mobile Cloud
6.3 Automatic Cloud Computing. .............. June-19,--------r-ii Marks 8
6.4 Multimedia Cloud
6.5 Energy Aware Cloud Computing ........... June-19,-- «csi Marks 8
6.6 Jungle Computing
6.7 Docker LL. June-19, Dec.-19, -------- Marks 8
6.8 Introduction to DevOps
6.9 IOT and Cloud Convergence
6.10 Multiple Choice Questions
(6-1)
Cloud Computing 6-2 Advanced Techniques in Cloud Computing
Ubuntu Linux-based OS, which has a JeOS version. Other JeOS platforms include
Novell's SUSE Linux Enterprise JeOS, Oracle’s Enterprise Linux JeOS, and Sun's
OpenSolaris JeOS.
me
Mobile cloud applications —_— Ry —
move the computing power A @ -—
and data storage away from
ne
mobile phones and into the
Smart mobile Wireless network Computational
cloud, bringing applications device technology cloud
and mobile computing to not
just smart phone users but a Fig. 6.2.1 Block diagram of mobile cloud
much broader range of mobile
subscribers.
Mobile cloud applications move the computing power and data storage away from
the mobile devices and into powerful and centralized computing platforms located
in clouds, which are then accessed over the wireless connection based on a thin
native client.
Mobile devices face many resource challenges (battery life, storage, bandwidth
etc.).
Cloud computing offers advantages to users by allowing them to use
infrastructure, platforms and software by cloud providers at low cost and
elastically in an on-demand fashion.
Mobile cloud computing provides mobile users with data storage and processing
services in clouds, obviating the need to have a powerful device configuration (e.g.
Mobile networkA
R Hh
|}
cloud service providers
ol
Cloud computing
Mobile network B
Hh
|}
ME
i point
| Contral Servers
et A
Application
processors
Internet service servers
providers (ISPs)
Mobile | BTS
devices 1
T T
1 Application service providers
In mobile cloud computing mobile network and cloud computing are combined,
thereby providing an optimal services for mobile clients.
Cloud computing exists when tasks and data are kept on individual devices.
Applications run on a remote server and then sent to the client.
Here the mobile devices are connected to the mobile networks through the base
stations; they will establish and control the connections (air interface) and
functional interfaces between the mobile networks and mobile devices.
Mobile users send service requests to the cloud through a web browser or desktop
application. The informations are transmitted to the central processors that are
connected to the servers providing mobile network services.
Here, services like AAA (Authentication, Authorization and Accounting) can be
provided to the users based on Home Agent (HA) and subscriber's data stored in
databases
Mobile devices are connected to the mobile networks via base stations that
establish and control the connections and functional interfaces between the
networks and mobile devices.
Mobile users’ requests and information are transmitted to the central processors
that are connected to servers providing mobile network services.
The subscribers’ requests are delivered to a cloud through the Internet.
In the cloud, cloud controllers process the requests to provide mobile users with
the corresponding cloud services.
Disadvantages :
1. Must send the program states (data) to the cloud server
2. Network latency can lead to execution delay.
- Ny g ~ 73
» Ax = | :
3 < Saa$ applications portal 7 my
EX
6.3.1 | CometCloud
Application
Master/Worker/BOT
Programming MaReduce/
layer ) oo Task Workflow Hadoop
Scheduling Monitoring :
consistency
Clustering/ —
Service Anotomy detection Coordination Publish/Subscribe
layer
Discovery Event Messaging
Self-organizing layer
Data center/Grid/Cloud
1. Describe cloudlets for mobile cloud computing with neat diagram and differentiate between
cloudlets and clouds. SPPU : June-19, Fnd Sem, Marks 8
XA Multimedia Cloud
e Due to the invention of cloud computing, nowadays users can easily access the
multimedia content over the internet at any time. User can efficiently store the
multimedia content of any type and of any size in the cloud after subscribing it
with no difficulties.
¢ Not only storing the media content like Audio, Video and Image, but can process
them within the cloud since the computation time for processing media data is
more in complex hardware.
e After processing the processed data can be easily received from the cloud through
a client without any need of installing complex hardware.
CA = 0 0 0 Media cloud
2
Storage
E—
CPU
Cloud media
Authoring/editing Sharing/streaming
service service
MSPs
1
oP0 K——> Load balancer
Not only processing, but the media content can be shared between clouds using
the streaming protocols like TCP/IP, UDP, RTP, HTTP etc.
Streaming of media content involves, loading or buffering media data, coding,
mixing, rating and rendering over the service providers.
Other profiling, packetizing, tokenizing of media contents will be done by the
cloud based on the streaming protocols used and it will be streamed to the client
system.
Cloud media technology offers number of key benefits to its service providers as
well as the users through increased implementation time, efficient data storage
capacity, less computation and cost.
It created a striking impact in the multimedia content processing like editing,
storing, encrypting and decrypting, gaming, streaming, compressing etc.
6.4.1 Hla"
Today, consumers watch video on a variety of connected devices. New
Over-The-Top (OTT) providers such as Netflix are offering direct-to-consumer
services with low prices, advanced user interfaces and easy access to multi-screen
video.
Changing usage patterns brought on by subscriber desire to watch content at the
time, location and on the device of their choosing are increasing content
distribution costs.
Pay TV providers are particularly susceptible to these trends and need to adapt
their traditional TV delivery architectures to offer innovative services that attract
and retain customers.
The traditional Set-Top Box (STB) will disappear. The functions of today’s STB
hardware will be carried out in the network and by the connected device itself,
eliminating the cost and complexity of managing home-based STBs.
Traffic will be all unicast. Over time, device format fragmentation, time-shifting
viewing habits and service personalization will erode broadcast and multicast
efficiencies.
Ultimately, every end user will be served with a unique stream. Services will be
deployed in the cloud.
Dedicated video platforms will migrate to cloud-based services, reducing costs and
accelerating time to market.
Operators will move from vertically integrated middleware stacks to more open
architectures with best-of-breed components.
Cloud DVR technology makes all TV content available on demand, on any device
and in any location.
There are several advantages of cloud-based services-the key benefits are that the
service is software-based, so one doesn’t need a physical location to run the
operations. As a result, real estate, infrastructure and manpower cost reduce
dramatically.
For instance, if a Bollywood film channel wants to check if it has a market in the
US, it can use the cloud-based technology to launch without adding to its cost by
leasing a satellite for a minimum of three to five years.
A product of ActiveVideo, a Silicon Valley software company, CloudTV is
available on more than 15 million devices.
Techniques Sub-types
e With clusters, grids and clouds thus being equipped with multi-core processors
and many-core ‘add-ons’, systems available to scientists are becoming increasingly
hard to program and use.
e Despite the fact that the programming and efficient use of many-cores is known to
be hard, this is not the only problem. With the increasing heterogeneity of the
underlying hardware, the efficient mapping of computational problems onto the
‘bare metal’ has become vastly more complex. Now more than ever, programmers
must be aware of the potential for parallelism at all levels of granularity.
Process Simplification
e Docker can simplify both workflows and communication, and that usually starts
with the deployment story.
e Fig. 6.7.1 shows workflow with and without docker.
1
Development team Operations team
i
Request resources : Resources provisioned
I
' i
Script the deployment Tweak the deployment
T
v i1
T
Application is deployed
. Go to step 5 and 6
. The application is deployed.
00
Build image
1
Deploy
Application is deployed
e Docker is increasingly well supported, with the majority of the large public clouds.
For example, Docker runs on AWS Elastic Beanstalk, Google AppEngine, IBM
Cloud, Microsoft Azure, etc.
e Google's Eric Brewer announced that Google would be supporting Docker as its
primary internal container format. Rather than just being good PR for these
companies, what this means for the Docker community is that there is starting to
be a lot of money backing the stability and success of the Docker platform.
Docker has traditionally been developed on the Ubuntu Linux distribution, but
today most Linux distributions and other major operating systems are now
supported where possible.
Architecture
The fundamental architecture of Docker is a simple client - server model, with
only one executable that acts as both components, depending on how you invoke
the docker command.
Underneath those simple exteriors, Docker heavily leverages kernel mechanisms
such as IPTABLES, virtual bridging, cgroups, namespaces, and various filesystem
drivers.
Fig. 6.7.3 shows docker architecture.
IF [i
ol [gf [gf 180—
= ——
‘a
| | ‘a al "2
IE
—
1 1
apes 1
= I =| ol 1°] 19 I nede ]
Client Docker daemon 1 TS 1
[Festos 1 i MMT
Server Docker host Docker registry
It consists of two parts : The client and the server. Registry is one more
components which stores docker images and metadata about those images.
Docker Engine is a client-server based application with following components -
1. A server which is a continuously running service called a daemon process.
2. A REST API which interfaces the programs to use talk with the daemon and
give instruct it what to do.
3. A command line interface client.
Docker client is the primary service using which docker users communicate with
the docker. When we use commands "docker run" the client sends these
commands to dockerd, which execute them out.
Docker server
on Linux host or Virtual machine
Docker registry
on Linux host, VM or SaaS
Docker client
on local workstation
¢ The command used by docker depend on docker AP. In docker client can interact
more than one daemon process.
e The docker images are building the block of docker or docker image is a read-only
template with instructions to create a docker container. Docker images are the
most build part of docker life cycle
e The server does the ongoing work of running and managing your containers, and
you use the client to tell the server what to do.
¢ The docker daemon can run on any number of servers in the infrastructure, and a
single client can address any number of servers.
e Clients drive all of the communication, but docker servers can talk directly to
image registries when told to do so by the client.
¢ Clients are responsible for directing servers what to do and servers focus on
hosting containerized applications.
e Docker registry keeps docker images. We can run our private registry.
¢ When we run the docker pull and docker run commands, the required images are
pulled from our configured registry directory.
e Using docker push command, the image can be uploaded to our configured
registry directory.
TT eer TN
(Neer
Containers Containers
Control plane =
User interface POD 1 POD 2
Controller manager
Containers Containers
Kubectl
— POD 1 POD 2
( Docker engine |
1. Explain the client server architecture of docker ? What are network ports and unix sockets ?
SPPU : June-19, End Sem, Marks 8
2. Explain docker with respect to process simplification. Broad support and Adoption,
architecture. SPPU : June-19, End Sem, Marks 8
xl Introduction to DevOps
DevOps is a collaboration between Development and IT Operations to make
software production and deployment in an automated and repeatable way.
DevOps helps to increase the organization's speed to deliver software applications
and services. The word 'DevOps' is a combination of two words, Development’
and 'Operations.’
Developers are willing to push out software faster and faster. Operations, on the
other hand, knows that rapid-fire changes without proper safeguards could
destabilize the system, which goes directly against their charter.
DevOps should be used for large distributed applications such as eCommerce sites
or applications hosted on a cloud platform.
DevOps bridges the gap between development and operations, creating significant
efficiencies across the development and deployment of software. DevOps includes
a strong emphasis on automation, helping reduce the overall number of errors.
pa
Thermostat
vl J
a rr
Smart living
| ——.
Fig. 6.9.1 Smart home
e In automatic light control system, Light Dependent Resistor (LDR) sensor is used
to detect bright /medium /dim /dark conditions.
While during mist and foggy weathers require the highest possible lighting level,
as the visibility reaches its lowest. On snowy weather it might require an
intermediate level between clear and foggy.
During night it requires high lighting levels, while at day it needs just fade level
to provide guidance or turn off if the weather is clear. The lighting concentration
in the yard is affected by the above conditions.
Today, users of IoT devices can evaluate engine performance, control air
temperature, and measure physical health indicators with only a few clicks.
Conventional perceptions of the automotive industry are radically changing with
IoT development. Predictive maintenance, Wi-Fi capabilities powered by
3G/4G/5G functionality, Car2Car connectivity, and advanced fleet management
are only a few examples of how IoT-based solutions are shaping the new
automotive age.
The automobile industry is one of the fastest-growing markets for IoT-based
solutions. The number of installed connectivity units in vehicles is likely to
increase by 67 % between 2018 and 2020.
Predictive maintenance technology is based on the use of IoT connectivity tools
that collect data on the performance of different parts, transfer that data to the
cloud in real time, and evaluate the risks of potential malfunction of a car's
I=T=T=1= « = » I=T=T=1=
—_— Low battery — —_— [4]
predicted 5
Battery Alert sent Issue
conditions avoided
Predictive maintenance can facilitate vehicle use by both private owners and
dealerships with large fleets of vehicles. It enables end-users to get the right
information in advance. With IoT connectivity tools, you can forget about
unplanned stops or breakdowns during the ride.
3. Medical resource shoriness : There are not enough medical resources for the
population. For example, there are fewer doctors and high-level healthcare
institutions but more patients.
"Low" usage of community health service centers. In contrast with community
health service centers, people prefer the high-level healthcare institutions. This
results in the low usage of community service centers.
Bad health habits : The citizens have some bad health habits that contribute to
poor health, for instance, smoking and no sport.
Lack of information sharing : Hospitals are not sharing enough information.
This leads to the following two problems at least. First, the health information
records of patients cannot be queried. Second, there is lack of medical
cooperation between hospitals.
e The links between the many applications in health monitoring are :
1. Applications require the gathering of data from sensors
2. Applications must support user interfaces and displays
3. Applications require network connectivity for access to infrastructural services
4 . Applications have in-use requirements such as low power, robustness,
durability, accuracy and reliability.
e Connected medical devices and associated IoT technologies will primarily be used
to achieve the following capabilities :
1. Access real time visibility of the patient's condition, his/her activities, context
and physiological parameters
Monitor compliance to prescribed treatment, diet and exercise regimes
3. Provide feedback and cues to patients, family members, doctors and caregivers
in order to implement corrective action
Leverage high performance computing for real time feedback and use
evidence-based medicine for better patient outcome.
a| Multitenancy b| Elasticity
Cc Homogeneous cloud platform d| On-demand service
Q.2 are a set of agreements that are signed between the user and service
providers.
Q1 c Q2 a Q3 a Q4
Q.5 d Q.6 c
aaa
Notes
Explain data intensive technologies for cloud computing. [Refer section 2.6.5] @
Explain cloud data stores. Disucss data store types. [Refer section 2.4] @
OR
Q4 a) What is cloud data management interface ? [Refer section 2.6.2] 3)
What is cloud provisioning ? Explain its benefits and characteristics. Also explain its types.
[Refer section 2.2.3] (v4)
M-1)
Cloud Computing M-2 Solved Model Question Papers
Q.1 a) What is server virtualization ? Explain requirement of server virtualization. List advantages
and disadvantages of server virtualization. [Refer section 3.3.2] ®
b) Explain full and para virtualization with examples. [Refer sections 3.3.4 and 3.4] (10)
OR
Q.2 a) Explain in brief virtual clusters and resource management. [Refer section 3.6] 5
b) Explain different levels of virtualization implementation with neat diagram. Also give
example of each. [Refer section 3.3] 5]
b) Explain steps to configure server for EC2 ? [Refer section 4.2] (0)
OR
Q.4 a) Explain the steps to create an Amazon S3 bucket and managing associated objects ?
[Refer section 4.2.2] @®
b) Discuss Google App engine. [Refer section 4.7] ©
Q.5 a) What is risk management ? Explain risk in cloud computing. [Refer section 5.1] (3)
c) How secure cloud software testing is performed? Discuss types of cloud testing.
[Refer section 5.9] (3)
Q.7 a) Explain traditional as well as docker deployment workflow, Client server architecture of
Docker ? [Refer section 6.7] ©
b) Describe cloudlets for mobile cloud computing with neat diagram and differentiate between
cloudlets and clouds. [Refer section 6.2] (3)
OR
Q.8 a) What is energy aware cloud computing ? Explain in detail. [Refer section 6.5] ®)
b) What is automatic cloud computing ? Discuss briefly CometCloud. [Refer section 6.3] (9)
aaa
2. Data Science and Big Data Analytics (. a. bhotre, or. kalpana v. Metre)
3. Artificial Intelligence {Anamitra Deshmukh-Nimbalkar, Dr. Vaishali P. Vikhe)
Elective Subjects
PAPER SOLUTIONS
m
© Covers Entire Syllabus e Question Answer Format ¢ Exact Answers & Solutions
© Important Points to Remember e Important Formulae
For All © Chapterwise Solved University Questions Last 10 Years Solved Papers
Branches
... Available at all Leading Booksellers ...