Data Analytics (IT-3006)
Kalinga Institute of Industrial Technology
Deemed to be University
Bhubaneswar-751024
School of Computer Engineering
Strictly for internal circulation (within KIIT) and reference only. Not for outside circulation without permission
3 Credit Lecture Note
Importance of the Course
2
The data analytics is indeed a revolution in the field of information technology.
The use of data analytics by the companies is enhancing every year and the primary
focus of the companies is on customers.
Many organizations are actively looking out for the right talent to analyze vast amounts
of data.
Following four perspectives leads to importance of data analytics.
Business Data
Data Science
Analytic
s Real-time
Job Market Usability
School of Computer Engineering
Why Learn Data Analytics?
3
A priority for top organizations.
Gain problem solving skills.
High demand
Increasing job opportunities.
Increasing pay.
Various job titles from which to choose (Metrics and Analytics
Specialist, Data Analyst, Big Data Engineer, Data Analytics Consultant)
Analytics is everywhere.
It's only becoming more important.
It represents perfect freelancing opportunities.
Develop new revenue streams.
School of Computer Engineering
Course Contents
4
Sr # Major and Detailed Coverage Area Hr
s
1 Introduction to Big Data 9
Introduction to Data, Big Data Characteristics, Types of Big Data, Challenges of Traditional,
Systems, Web Data, Evolution of Analytic Scalability, OLTP, MPP, Grid Computing, Cloud
Computing, Fault Tolerance, Analytic Processes and Tools, Analysis Versus Reporting, Statistical
Concepts, Types of Analytics.
2 Data Analysis 12
Introduction to Data Analysis, Importance of Data Analysis, Data Analytics Applications,
Regression Modelling Techniques: Linear Regression, Multiple Linear Regression, Non Linear
Regression, Logistic Regression, Bayesian Modelling, Basian Networks, Support Vector
Machines, Time Series Analysis, Rule Induction, Sequential Cover Algorithm.
3 Mining Data Streams 10
Introduction to Mining Data Streams, Data Stream Management Systems, Data Stream Mining,
Examples of Data Stream Applications, Stream Queries, Issues in Data Stream Query, Processing,
Sampling in Data Streams, Filtering Streams, Counting Distinct Elements in a Stream, Estimating
Moments, Querying on Windows − Counting Ones in a Window, Decaying Windows, Real-Time
Analytics Platform (RTAP).
School of Computer Engineering
Course Contents continue…
5
Sr # Major and Detailed Coverage Area Hrs
4 Frequent Itemsets and Clustering 10
Introduction to Frequent Itemsets, Market-Basket Model, Algorithm for Finding
Frequent Itemsets, Association Rule Mining, Apriori Algorithm, Introduction to
Clustering, Overview of Clustering Techniques, Hierarchical Clustering, Partitioning
Methods, K- Means Algorithm, Clustering High-Dimensional Data.
5 Frameworks and Visualization 8
Introduction to framework and Visualization, Introduction to Hadoop, Core
Components of Hadoop, Hadoop Ecosystem, Physical Architecture, Hadoop
Limitations, Hive, MapReduce and The New Software Stack, MapReduce, Algorithms
Using MapReduce, NOSQL, NoSQL Business Drivers, NoSQL Case Studies, NoSQL
Data Architectural Patterns, Variations of NoSQL, Architectural Patterns, Using
NoSQL to Manage Big Data, Visualizations
School of Computer Engineering
Course Outcome
6
CO # CO Unit
CO1 Understand and classify the characteristics, concepts and Introduction to
principles of big data. Big Data
CO2 Apply the data analytics techniques and models. Data Analysis
CO3 Implement and analyze the data analysis techniques for Mining Data
mining data streams. Streams
CO4 Examine the techniques of clustering and frequent item Frequent Itemsets
sets. and Clustering
CO5 Analyze and evaluate the framework and visualization for Frameworks and
big data analytics. Visualization
CO6 Formulate the concepts, principles and techniques focusing Applications of all
on the applications to industry and real world experience. units
Prerequisites
NIL
School of Computer Engineering
Books
7
Textbook
Data Analytics, Radha Shankarmani,M. Vijayalaxmi, Wiley India Private Limited,
ISBN: 9788126560639.
Reference Books
Data Science and Big Data Analytics: Discovering, Analyzing, Visualizing and
Presenting Data by EMC Education Services (Editor), Wiley, 2014
Bill Franks, Taming the Big Data Tidal Wave: Finding Opportunities in Huge Data
Streams with advanced analystics, John Wiley & sons, 2012.
Glenn J. Myatt, Making Sense of Data, John Wiley & Sons, 2007 Pete Warden, Big
Data Glossary,O’Reilly, 2011.
Jiawei Han, MichelineKamber “Data Mining Concepts and Techniques”, Second
Edition, Elsevier, Reprinted 2008.
Stephan Kudyba, Thomas H. Davenport, Big Data, Mining, and Analytics, Components
of Strategic Decision Making, CRC Press, Taylor & Francis Group. 2014
Big Data, Black Book, DT Editorial Services, Dreamtech Press, 2015
School of Computer Engineering
Evaluation
8
Grading:
Internal assessment – 30 marks
2 quizzes = 2.5 X 2 = 5 marks
5 group assignments = 2 X 5 = 10 marks
Class participation = 5 marks
Mini Project = 10 marks
Mid-Term exam - 20 marks
End-Term exam - 50 marks
?
School of Computer Engineering
Data
9
A representation of information, knowledge, facts, concepts or instructions
which are being prepared or have been prepared in a formalized manner.
Data is either intended to be processed, is being processed, or has been
processed.
It can be in any form stored internally in a computer system or computer
network or in a person’s mind.
Since the mid-1900s, people have used the word data to mean computer
information that is transmitted or stored.
Data is the plural of datum (a Latin word meaning something given), a single
piece of information. In practice, however, people use data as both the
singular and plural form of the word.
It must be interpreted, by a human or machine to derive meaning.
It is presents in homogeneous sources as well as heterogeneous sources.
The need of the hour is to understand, manage, process, and take the data
for analysis to draw valuable insights.
Data Information Knowledge Actionable Insights
School of Computer Engineering
Importance of Data
10
The ability to analyze and act on data is increasingly important to
businesses. It might be part of a study helping to cure a disease, boost a
company’s revenue, understand and interpret market trends, study
customer behavior and take financial decisions.
The pace of change requires companies to be able to react quickly to
changing demands from customers and environmental conditions. Although
prompt action may be required, decisions are increasingly complex as
companies compete in a global marketplace.
Managers may need to understand high volumes of data before they can
make the necessary decisions.
Relevant data creates strong strategies - Opinions can turn into great
hypotheses, and those hypotheses are just the first step in creating a strong
strategy. It can look something like this: “Based on X, I believe Y, which will
result in Z”.
Relevant data strengthens internal teams.
Relevant data quantifies the purpose of the work.
School of Computer Engineering
Characteristics of Data
11
Deals with the structure of the
data i.e. source, the granularity,
the type, nature whether static Composition
or real-time streaming
Deals with the state of the data
i.e. usability for analysis, does Condition Data
it require cleaning for further
enhancement and enrichment?
Deals with “where it has been Context
generated”, “ why was this
generated”, “how sensitive is
this”, “what are the associated
events” and so on.
School of Computer Engineering
Human vs. Machine Readable data
12
Human-readable refers to information that only humans can interpret and study,
such as an image or the meaning of a block of text. If it requires a person to
interpret it, that information is human-readable.
Machine-readable refers to information that computer programs can process. A
program is a set of instructions for manipulating data. Such data can be
automatically read and processed by a computer, such as CSV, JSON, XML, etc.
Non-digital material (for example printed or hand-written documents) is by its non-
digital nature not machine-readable. But even digital material need not be machine-
readable. For example, a PDF document containing tables of data. These are
definitely digital but are not machine-readable because a computer would struggle
to access the tabular information - even though they are very human readable. The
equivalent tables in a format such as a spreadsheet would be machine readable.
Another example scans (photographs) of text are not machine-readable (but are
human readable!) but the equivalent text in a format such as a simple ASCII text file
can machine readable and processable.
School of Computer Engineering
Classification of Digital Data
13
Digital data is classified into the following categories:
Structured data
Semi-structured data
Unstructured data
Approximate percentage distribution of digital data
School of Computer Engineering
Structured Data
14
It is defined as the data that has a well-defined repeating pattern and this
pattern makes it easier for any program to sort, read, and process the data.
This is data is in an organized form (e.g., in rows and columns) and can be easily
used by a computer program.
Relationships exist between entities of data.
Structured data:
Organize data in a pre-defined format.
Is stored in a tabular form.
Is the data that resides in a fixed fields within a record of file.
Is formatted data that has entities and their attributes mapped.
Is used to query and report against predetermined data types.
Sources:
Multidimension
Relational
al
database
Structured data databases
Legacy
Flat files
databases
School of Computer Engineering
Ease with Structured Data
15
Insert/Update/ DML operations provide the required ease with data
input, storage, access, process , analysis etc.
Delete
Encryption and tokenization solution to warrant the
security of information throughout life cycle.
Security Organization able to retain control and maintain
compliance adherence by ensuring that only authorized
are able to decrypt and view sensitive information.
Indexing speed up the data retrieval operation at the
Structured data Indexing cost of additional writes and storage space, but the
benefits that ensure in search operation are worth the
additional writes and storage spaces.
The storage and processing capabilities of the traditional
Scalability DBMS can be easily be scaled up by increasing the
horsepower of the database server.
Transaction RDBMS has support of ACID properties of transaction
to ensure accuracy, completeness and data integrity.
Processing
School of Computer Engineering
Semi-structured Data
16
Semi-structured data, also known as having a schema-less or self-describing
structure, refers to a form which does not conform to a data model as in
relational database but has some structure.
In other words, data is stored inconsistently in rows and columns of a database.
However, it is not in a form which can be used easily by a computer program.
Example, emails, XML, markup languages like HTML, etc. Metadata for this data
is available but is not sufficient.
Sources:
Web data in
the form of XML
cookies Semi-structured
data
Other Markup JSON
languages
School of Computer Engineering
XML, JSON, BSON format
17
Source (XML & JSON): http://sqllearnergroups.blogspot.com/2014/03/how-to-get-json-format-through-sql.html
Source (JSON & BSON): http://www.expert-php.fr/mongodb-bson/
School of Computer Engineering
Characteristics of Semi-structured Data
18
Inconsistent
Structure
Self-describing
(level/value pair)
Other schema
Semi-structured information is
data blended with data
values
Data objects may
have different
attributes not known
beforehand
School of Computer Engineering
Unstructured Data
19
Unstructured data is a set of data that might or might not have any logical or
repeating patterns and is not recognized in a pre-defined manner.
About 80 percent of enterprise data consists of unstructured content.
Unstructured data:
Typically consists of metadata i.e. additional information related to data.
Comprises of inconsistent data such as data obtained from files, social
media websites, satellites etc
Consists of data in different formats such as e-mails, text, audio, video, or
images.
Sources: Body of email
Chats, Text
Text both
messages
internal and
external to org.
Mobile data
Unstructured data
Social Media Images, audios,
data videos
School of Computer Engineering
Challenges associated with Unstructured data
20
Working with unstructured data poses certain challenges, which are as follows:
Identifying the unstructured data that can be processed.
Sorting, organizing, and arranging unstructured data indifferent sets and
formats.
Combining and linking unstructured data in a more structured format to derive
any logical conclusions out of the available information.
Costing in terms of storage space and human resources need to deal with the
exponential growth of unstructured data.
Data Analysis of Unstructured Data
The complexity of unstructured data lies within the language that created it. Human
language is quite different from the language used by machines, which prefer
structured information. Unstructured data analysis is referred to the process of
analyzing data objects that doesn’t follow a predefine data model and/or is
unorganized. It is the analysis of any data that is stored over time within an
organizational data repository without any intent for its orchestration, pattern or
categorization.
School of Computer Engineering
Dealing with Unstructured data
21
Data Mining (DM)
Natural Language Processing (NLP)
Dealing with
Unstructured data Text Analytics (TA)
Noisy Text Analytics
Note: Refer to Appendix for further details.
School of Computer Engineering
Definition of Big Data
22
Big Data is high-volume, high-velocity,
and high-variety information assets that
High-volume demand cost effective, innovative forms
High-velocity
High-variety of information processing for enhanced
insight and decision making.
Source: Gartner IT Glossary
Cost-effective,
innovative
forms of
information
processing
Enhanced
insight &
decision
making
School of Computer Engineering
What is Big Data?
23
Think of following:
Every second, there are around 822 tweets on Twitter.
Every minutes, nearly 510 comments are posted, 293 K statuses are updated,
and 136K photos are uploaded in Facebook.
Every hour, Walmart, a global discount departmental store chain, handles more
than 1 million customer transactions.
Everyday, consumers make around 11.5 million payments by using PayPal.
In the digital world, data is increasing rapidly because of the ever increasing use of
the internet, sensors, and heavy machines at a very high rate. The sheer volume,
variety, velocity, and veracity of such data is signified the term ‘Big Data’.
Semi- Big
Structured Unstructure
structured Data
Data d Data
Data
School of Computer Engineering
Elements of Big Data
24
In most big data circles, these are called the four V’s: volume, variety, velocity, and veracity.
(One might consider a fifth V, value.)
Volume - refers to the incredible amounts of data generated each second from social media,
cell phones, cars, credit cards, M2M sensors, photographs, video, etc. The vast amounts of
data have become so large in fact it can no longer store and perform data analysis using
traditional database technology. So using distributed systems, where parts of the data is
stored in different locations and brought together by software.
Variety - defined as the different types of data the digital system now use. Data today looks
very different than data from the past. New and innovative big data technology is now
allowing structured and unstructured data to be harvested, stored, and used
simultaneously.
Velocity - refers to the speed at which vast amounts of data are being generated, collected
and analyzed. Every second of every day data is increasing. Not only must it be analyzed,
but the speed of transmission, and access to the data must also remain instantaneous to
allow for real-time access. Big data technology allows to analyze the data while it is being
generated, without ever putting it into databases.
Veracity - is the quality or trustworthiness of the data. Just how accurate is all this data?
For example, think about all the Twitter posts with hash tags, abbreviations, typos, etc., and
the reliability and accuracy of all that content.
School of Computer Engineering
Elements of Big Data cont’d
25
Value - refers to the ability to transform a tsunami of data into business. Having endless
amounts of data is one thing, but unless it can be turned into value it is useless.
Refer to Appendix
for data volumes
School of Computer Engineering
Why Big Data?
26
More data for analysis will result into greater analytical accuracy and greater
confidence in the decisions based on the analytical findings. This would entail a greater
positive impact in terms of enhancing operational efficiencies, reducing cost and time,
and innovating on new products, new services and optimizing existing services.
More data
More accurate analysis
Greater confidence in decision making
Greater operational efficiencies, cost
reduction, time reduction, new
product development, and optimized
offering etc.
School of Computer Engineering
Challenges of Traditional Systems
27
The main challenge in the traditional approach for computing systems to manage
‘Big Data’ because of immense speed and volume at which it is generated. Some of
the challenges are:
Traditional approach cannot work on unstructured data efficiently.
Traditional approach is built on top of the relational data model, relationships
between the subjects of interests have been created inside the system and the
analysis is done based on them. This approach will not adequate for big data.
Traditional approach is batch oriented and need to wait for nightly ETL
(extract, transform and load) and transformation jobs to complete before
the required insight is obtained.
Traditional data management, warehousing, and analysis systems fizzle to
analyze this type of data. Due to it’s complexity, big data is processed with
parallelism. Parallelism in a traditional system is achieved through costly
hardware like MPP (Massively Parallel Processing) systems.
Inadequate support of aggregated summaries of data.
School of Computer Engineering
Challenges of Traditional Systems cont’d
28
Other challenges can be categorized as:
Data Challenges:
Volume, velocity, veracity, variety
Data discovery and comprehensiveness
Scalability
Process challenges
Capturing Data
Aligning data from different sources
Transforming data into suitable form for data analysis
Modeling data(Mathematically, simulation)
Management Challenges:
Security
Privacy
Governance
Ethical issues
School of Computer Engineering
Web Data
29
It refers to the data that is publicly available on the
web sites.
The web data has documents in pdf, doc, plain text as
well as images, music, and videos.
The most widely used and best‐known source of big
data today is the detailed data collected from web
sites.
The data is unstructured and inappropriate for access
by software application, and hence is converted to
either semi-structured or structured format that is
well suited for both humans and machines.
School of Computer Engineering
Distributed vs. Parallel Computing
30
Parallel Computing Distributed Computing
Shared memory system Distributed memory system
Multiple processors share a Autonomous computer nodes
single bus and memory unit connected via network
Processor is order of Tbps Processor is order of Gbps
Limited Scalability Better scalability and cheaper
Distributed computing in local
network (called cluster
computing). Distributed
computing in wide-area network
(grid computing)
School of Computer Engineering
EDW, OLTP, MPP
31
Enterprise Data Warehouse: An enterprise data warehouse (EDW) is a database,
or collection of databases, that centralizes a business's information from multiple
sources and applications, and makes it available for analytics and use across the
organization. EDWs can be housed in an on-premise server or in the cloud. The data
stored in this type of digital warehouse can be one of a business’s most valuable
assets, as it represents much of what is known about the business, its employees, its
customers, and more.
Online Transactional Processing (OLTP): It is a category of data processing that is
focused on transaction-oriented tasks. OLTP typically involves inserting, updating,
and/or deleting small amounts of data in a database. OLTP mainly deals with large
numbers of transactions by a large number of users.
Massively Parallel Processing (MPP): It is a storage structure designed to handle
the coordinated processing of program operations by multiple processors. This
coordinated processing can work on different parts of a program, with each
processor using its own operating system and memory. This allows MPP databases
to handle massive amounts of data and provide much faster analytics based on large
datasets.
School of Computer Engineering
Hadoop
32
Hadoop is an open-source framework that allows to store and process big data in a
distributed environment across clusters of computers using simple programming
models.
It is designed to scale up from single servers to thousands of machines, each offering
local computation and storage.
It provides massive storage for any kind of data, enormous processing power and
the ability to handle virtually limitless concurrent tasks or jobs.
Importance:
Ability to store and process huge amounts of any kind of data, quickly.
Computing power: It’s distributed computing model processes big data fast.
Fault tolerance: Data and application processing are protected against
hardware failure.
Flexibility: Unlike traditional relational databases, preprocess of data does not
require before storing it.
Low cost: The open-source framework is free and uses commodity hardware to
store large quantities of data.
Scalability: System can easily grow to handle more data simply by adding
nodes. Little administration is required.
School of Computer Engineering
Evolution of Analytics Scalability
33
As the amount of data organizations process continue to
increase, the world of big data requires new levels of
scalability. Organizations need to update the technology to
provide a higher level of scalability.
Luckily, there are multiple technologies available that address
different aspects of the process of taming big data and making
use of it in analytic processes.
The technologies are:
MPP (massively parallel processing)
Cloud computing
Grid computing
MapReduce
School of Computer Engineering
Traditional Analytics Architecture
34
Database 1
Analytic Server
Database 2
Extract
Database 3
The heavy processing occurs in the analytic environment. This
Database n may even a PC.
School of Computer Engineering
Modern In-Database Analytics Architecture
35
Database 1
Analytic Server
Database 2
Submit
Consolidate
Request
Database 3 Enterprise Data
Warehouse (EDW)
Database n
In an in-database environment, the processing stays in the database where the data
has been consolidated. The user’s machine just submits the request; it doesn’t do
heavy lifting.
School of Computer Engineering
MPP Analytics Architecture
36
Massively parallel processing (MPP) database systems is the most mature, proven, and
widely deployed mechanism for storing and analyzing large amounts of data. An MPP
database spreads data out into independent pieces managed by independent storage
and central processing unit (CPU) resources. Conceptually, it is like having pieces of
data loaded onto multiple network connected personal computers around a house.
The data in an MPP system gets split across a variety of disks managed by a variety of
CPUs spread across a number of servers.
Single overloaded server
In stead of single
overloaded database, an
MPP database breaks the
data into independent
chunks with independent
Multiple lightly loaded server
disk and CPU.
School of Computer Engineering
MPP Database Example
37
100-gigabyte 100-gigabyte 100-gigabyte 100-gigabyte 100-gigabyte
chunks chunks chunks chunks chunks
One-terabyte
table 100-gigabyte 100-gigabyte 100-gigabyte 100-gigabyte 100-gigabyte
chunks chunks chunks chunks chunks
A Traditional database will query
a one-terabyte table one row at time 10 simultaneous 100-gigabyte queries
MPP database is based on the principle of SHARE THE WORK!
A MPP database spreads data out across multiple sets of CPU and disk space. Think
logically about dozens or hundreds of personal computers each holding a small piece of a
large set of data. This allows much faster query execution, since many independent
smaller queries are running simultaneously instead of just one big query.
If more processing power and more speed are required, just bolt on additional capacity in
the form of additional processing units.
MPP systems build in redundancy to make recovery easy and have resource
management tools to manage the CPU and disk space.
School of Computer Engineering
MPP Database Example cont’d
38
An MPP system allows the different sets of CPU and disk to run the process concurrently
An MPP system
breaks the job into pieces
Single Threaded
Process ★ Parallel Process ★
School of Computer Engineering
OLTP vs. MPP vs. Hadoop
39
OLTP MPP
Examples: Oracle, DB2, SQL Server etc. Examples: Netezza, Teradata, Vertica etc.
It needs to read data from disk to memory before Takes the processing as close possible to the data,
start processing, so very fast in memory so less data movement
calculation.
It is good for smaller OLTP (transaction) It is good for batch processing. Some of the MPP
operations. It also maintains very high level of (Netezza, Vertica) overlooks integrity like
data integrity. enforcing unique key for the sake of batch
performance.
MPP Hadoop
Stores data in a matured internal structure. So data There are no such structured architecture for data
loading and data processing is efficient. stored on Hadoop. So, accessing and loading data
is not as efficient as conventional MPP systems.
It support only relational models. Support virtually any kind of data.
However the main objective of MPP and Hadoop is same, process data parallely near storage.
School of Computer Engineering
How to choose what?
40
OLTP Databases (Oracle,DB2, MySQL, MS SQL, Exadata):
Transaction based application
Smaller data warehouse
MPP (Netezza, Teradata, Vertica):
Bigger Data warehouse (may be having tables with size more than 4-5 TB)
Needs no or little pre-processing
Needs faster batch processing speed
In database analytics
Only Hadoop:
All data as heavily unstructured (documents, audio, video etc)
Need to process in batch
School of Computer Engineering
Grid Computing
41
Grid Computing can be defined as a network of computers working
together to perform a task that would rather be difficult for a single
machine.
The task that they work on may include analysing huge datasets or
simulating situations which require high computing power.
Computers on the network contribute resources like processing
power and storage capacity to the network.
Grid Computing is a subset of distributed computing, where a virtual
super computer comprises of machines on a network connected by
some bus, mostly Ethernet or sometimes the Internet.
It can also be seen as a form of parallel computing where instead of
many CPU cores on a single machine, it contains multiple cores
spread across various locations.
School of Computer Engineering
How Grid Computing works?
42
School of Computer Engineering
Cloud Computing
43
It is a internet-based computing and relies on sharing computing
resources on-demand rather than having local PCs and other devices.
It is the delivery of on-demand computing services - from
applications to storage and processing power over the internet and
on a pay-as-you-go basis.
It uses high-capacity networks, low-cost computers, and storage
devices and adopts hardware virtualization, service-oriented
architecture, and utility computing.
Rather than owning their own computing infrastructure or data
centers, companies can rent access to anything from applications to
storage from a cloud service provider and can scale up and scale
down as per their computing demands.
There are 3 types of cloud environment named public cloud, private
cloud and hybrid cloud.
School of Computer Engineering
Public Cloud
44
It is the most common type of cloud computing deployment.
The cloud resources (like servers and storage) are owned and
operated by a third-party cloud service provider and delivered over
the internet.
With a public cloud, all hardware, software and other supporting
infrastructure are owned and managed by the cloud provider.
In a public cloud, the same hardware, storage and network devices
are shared with other organizations or cloud “tenants,” and the
adopter access services and manage account using a web browser.
Public cloud deployments are frequently used to provide web-based
email, online office applications, storage and testing and
development environments.
Advantages of public clouds are lower costs, no maintenance, high
reliability etc.
School of Computer Engineering
Private Cloud
45
A private cloud consists of cloud computing resources used
exclusively by one business or organization.
The private cloud can be physically located at your organization's on-
site datacenter or it can be hosted by a third-party service provider.
The services and infrastructure are always maintained on a private
network and the hardware and software are dedicated solely to the
organisation.
It is often used by government agencies, financial institutions, any
other mid- to large-size organizations with business-critical
operations seeking enhanced control over their environment.
Advantages of private clouds are more flexibility, more control, and
more scalability etc.
School of Computer Engineering
Hybrid Cloud
46
A hybrid cloud combines on-premises infrastructure or a private
cloud with a public cloud.
It allow data and apps to move between the two environments.
Many organizations choose a hybrid cloud approach due to business
imperatives such as meeting regulatory and data sovereignty
requirements, taking full advantage of on-premises technology
investment or addressing low latency issues.
A hybrid cloud platform gives organizations many advantages—such
as greater flexibility, more deployment options, security, compliance
and getting more value from their existing infrastructure.
When computing and processing demand fluctuates, hybrid cloud
computing gives businesses the ability to seamlessly scale up their
on-premises infrastructure to the public cloud to handle any
overflow—without giving third-party datacenters access to the
entirety of their data.
School of Computer Engineering
Fault Tolerance
47
Fault tolerance refers to the ability of a system (computer, network, cloud cluster,
etc.) to continue operating without interruption when one or more of its
components fail.
The objective of creating a fault-tolerant system is to prevent disruptions arising
from a single point of failure, ensuring the high availability and business continuity
of mission-critical applications or systems.
Fault-tolerant systems use backup components that automatically take the place of
failed components, ensuring no loss of service. These include:
Hardware systems that are backed up by identical or equivalent systems. For
example, a server can be made fault tolerant by using an identical server
running in parallel, with all operations mirrored to the backup server.
Software systems that are backed up by other software instances. For example,
a database with customer information can be continuously replicated to
another machine. If the primary database goes down, operations can be
automatically redirected to the second database.
Power sources that are made fault tolerant using alternative sources. For
example, many organizations have power generators that can take over in case
main line electricity fails.
School of Computer Engineering
Analytic Processes and Tools
48
Self-Study from the book
Points to cover
Spreadsheets and Analytics Tool
Analytics Engine
CRM and Online Marketing Solutions
School of Computer Engineering
Analysis vs. Reporting
49
Reporting: The process of organizing data into informational summaries in
order to monitor how different areas of a business are performing.
Analysis: The process of exploring data and reports in order to extract
meaningful insights, which can be used to better understand and improve
business performance.
Difference b/w Reporting and Analysis:
Reporting translates raw data into information. Analysis transforms data
and information into insights.
Reporting helps companies to monitor their online business and be alerted
to when data falls outside of expected ranges. Good reporting should raise
questions about the business from its end users. The goal of analysis is to
answer questions by interpreting the data at a deeper level and providing
actionable recommendations.
In summary, reporting shows you what is happening while analysis focuses
on explaining why it is happening and what you can do about it.
School of Computer Engineering
Goal of Analysis and Reporting
50
Reporting uses data to track the performance of your business, while an
analysis uses data to answer strategic questions about your business. Though
they are distinct, reporting and analysis rely on each other. Reporting sheds
light on what questions to ask, and an analysis attempts to answer those
questions.
Simply put,
Data Reporting Reveals The Right Questions.
Data Analysis Helps Find Answers.
School of Computer Engineering
Data Analytics
51
Data analytics is the process of extracting useful information by analysing
different types of data sets. It is used to discover hidden patterns, outliers,
unearth trends, unknown co-relationship and other useful information for the
benefit of faster decision making.
There are 4 types of analytics:
School of Computer Engineering
Types of Analytics
52
Approach Explanation
Descriptive What’s happening in my business?
•Comprehensive, accurate and historical data
•Effective Visualisation
Diagnostic Why is it happening?
•Ability to drill-down to the root-cause
•Ability to isolate all confounding information
Predictive What’s likely to happen?
•Decisions are automated using algorithms and technology
•Historical patterns are being used to predict specific outcomes using
algorithms
Prescriptive What do I need to do?
•Recommended actions and strategies based on champion/challenger
strategy outcomes
•Applying advanced analytical algorithm to make specific
recommendations
School of Computer Engineering
Mapping of Big Data’s Vs to Analytics Focus
53
History data can be quite large. There might be a need to process huge amount of data many times a
day as it gets updated continuously. Therefore volume is mapped to history. Variety is pervasive.
Input data, insights, and decisions can span a variety of forms, hence it is mapped to all three. High
velocity data might have to be processed to help real time decision making and plays across
descriptive, predictive, and prescriptive analytics when they deal with present data. Predictive and
prescriptive analytics create data about the future. That data is uncertain, by nature and its veracity
is in doubt. Therefore veracity is mapped to prescriptive and predictive analytics when it deal with
future.
School of Computer Engineering
Big Data Analytics
54
Big data analytics is the process of extracting useful information by analysing different
types of big data sets. It is used to discover hidden patterns, outliers, unearth trends,
unknown co-relationship and other useful info for the benefit of faster decision making.
Big Data Application in different Industries
School of Computer Engineering
What is Big Data Analytics ?
55
Richer, deeper insights into
Better, faster decisions in Move code to data for
customers, partners and the
real-time greater speed and efficiency
business
Working with datasets
whose volume and variety Big Data
Competitive advantages
is beyond the storage and Analytics
capacity of typical DB
IT’s collaboration with Time-sensitive decisions
Technology enabled
business users and data made in near real time by
analytics
scientist processing real-time data
School of Computer Engineering
What is Big Data Analytics isn’t?
56
Only about Volume Just about technology Meant to replace RDBMS
Big Data
Analytics isn’t
“One-size-fit-all” traditional
Only used by huge online Meant to replace data
RDBMS built on shared
companies warehouse
disk and memory
School of Computer Engineering
Top challenges facing Big Data
57
1. Scale: Storage is one major concern that needs to be addressed to handle
the need for scaling rapidly and elastically. The need of the hour is a storage
that can best withstand the onslaught of large volume, velocity, and variety
of big data? Should scale vertically or horizontally?
2. Security: Most of the NoSQL (Not only SQL) big data platforms have poor
security mechanism (lack of proper authentication and authorization
mechanisms) when it comes to safeguarding big data.
3. Schema: Rigid schema have no place. The need of the hour is dynamic
schema and static (pre-defined) schemas are passed.
4. Data Quality: How to maintain data quality – data accuracy, completeness,
timeliness etc. Is the appropriate metadata in place?
5. Partition Tolerant: How to build partition tolerant systems that can take
care of both hardware and software failures?
6. Continuous availability: The question is how to provide 24/7 support
because almost all RDBMS and NoSQL big data platforms have a certain
amount of downtime built in.
School of Computer Engineering
Kind of Technologies to help meet the
challenges posed by Big Data
58
1. Cheap and abundant storage
2. Faster processors to help with quicker processing of
big data
3. Affordable open-source, distributed big data
platforms
4. Parallel processing, clustering, visualisation, large grid
environments, high connectivity, and high
throughputs rather than low latency
5. Cloud computing and other flexible resource
allocation agreements
School of Computer Engineering
Summary
59
Detailed Lessons
Introduction to Data, Big Data Characteristics, Types of Big Data, Challenges of
Traditional, Systems, Web Data, Evolution of Analytic Scalability, OLTP, MPP, Grid
Computing, Cloud Computing, Fault Tolerance, Analytic Processes and Tools, Analysis
Versus Reporting, Types of Analytics.
How was the journey?
School of Computer Engineering
60
School of Computer Engineering
Appendix
61
Data Mining: Data mining is the process of looking for hidden, valid, and
potentially useful patterns in huge data sets. Data Mining is all about
discovering unsuspected/previously unknown relationships amongst the
data. It is a multi-disciplinary skill that uses machine learning, statistics,
AI and database technology.
Natural Language Processing (NLP): NLP gives the machines the ability
to read, understand and derive meaning from human languages.
Text Analytics (TA): TA is the process of extracting meaning out of text.
For example, this can be analyzing text written by customers in a
customer survey, with the focus on finding common themes and trends.
The idea is to be able to examine the customer feedback to inform the
business on taking strategic action, in order to improve customer
experience.
Noisy text analytics: It is a process of information extraction whose goal
is to automatically extract structured or semi-structured information
from noisy unstructured text data.
School of Computer Engineering
Appendix cont… Example of Data Volumes
62
Unit Value Example
Kilobytes (KB) 1,000 bytes a paragraph of a text document
Megabytes (MB) 1,000 Kilobytes a small novel
Gigabytes (GB) 1,000 Megabytes Beethoven’s 5th Symphony
Terabytes (TB) 1,000 Gigabytes all the X-rays in a large hospital
Petabytes (PB) half the contents of all US academic research
1,000 Terabytes
libraries
Exabytes (EB) about one fifth of the words people have ever
1,000 Petabytes
spoken
Zettabytes (ZB) 1,000 Exabytes as much information as there are grains of sand on
all the world’s beaches
Yottabytes (YB) 1,000 Zettabytes as much information as there are atoms in 7,000
human bodies
School of Computer Engineering
Appendix cont… ETL
63
ETL is short for extract, transform, load, three database functions that are
combined into one tool to pull data out of one database/data sources and
place it into another database.
Extract is the process of reading data from a database. In this stage, the
data is collected, often from multiple and different types of sources.
Transform is the process of converting the extracted data from its
previous form into the form it needs to be in so that it can be placed into
another database. Transformation occurs by using rules or lookup tables
or by combining the data with other data.
Load is the process of writing the data into the target database.
School of Computer Engineering
Appendix cont… Utility Computing
64
Utility computing is the most trending IT service model. It provides on-
demand computing resources (computation, storage, and programming
services via API) and infrastructure based on the pay per use method. It
minimizes the associated costs and maximizes the efficient use of resources.
The advantage of utility computing is that it reduced the IT cost, provides
greater flexibility, and easier to manage.
The foundational concept is that users or businesses pay the providers of
utility computing for the amenities used – such as computing capabilities,
storage space and applications services. The customer is thus, absolved
from the responsibility of maintenance and management of the hardware.
Consequently, the financial layout is minimal for the organization.
Utility computing helps eliminate data redundancy, as huge volumes of
data are distributed across multiple servers or backend systems. The
client however, can access the data anytime and from anywhere.
School of Computer Engineering
Appendix cont… Multidimensional Databases
65
Multidimensional databases are used mostly for OLAP (online analytical
processing) and data warehousing. They can be used to show multiple
dimensions of data to users. A multidimensional database is created from
multiple relational databases. While relational databases allow users to access
data in the form of queries, the multidimensional databases allow users to ask
analytical questions related to business or market trends.
Example
The revenue costs for a company can be
understood and analyzed on the basis of
various factors like the company products,
the geographical locations of the company
offices, time to develop a product,
promotions done etc.
School of Computer Engineering