Data Analytics Unit 4
Data Analytics Unit 4
S
.R
Dr. Radhey Shyam
Dr
Professor
Unit-4 has been prepared and compiled by Dr. Radhey Shyam, with grateful acknowledgment to those who
made their course contents freely available or (Contributed directly or indirectly). Feel free to use this
,
12
study material for your own academic purposes. For any query, communication can be made through this
email : [email protected].
ril
Ap
CO 5 Describe the concept of R programming and implement analytics on Big data using R. K2,K3
S
Unit Topic Proposed
Lecture
.R
Introduction to Data Analytics: Sources and nature of data, classification of data
(structured, semi-structured, unstructured), characteristics of data, introduction to Big Data
platform, need of data analytics, evolution of analytic scalability, analytic process and
I tools, analysis vs reporting, modern data analytic tools, applications of data analytics. 08
Dr
Data Analytics Lifecycle: Need, key roles for successful analytic projects, various phases
of data analytics lifecycle – discovery, data preparation, model planning, model building,
communicating results, operationalization.
4/
Data Analysis: Regression modeling, multivariate analysis, Bayesian modeling, inference
and Bayesian networks, support vector and kernel methods, analysis of time series: linear
2
II systems analysis & nonlinear dynamics, rule induction, neural networks: learning and
20
08
generalisation, competitive learning, principal component analysis and neural networks,
fuzzy logic: extracting fuzzy models from data, fuzzy decision trees, stochastic search
methods.
,
Mining Data Streams: Introduction to streams concepts, stream data model and
12
Apriori algorithm, handling large data sets in main memory, limited pass algorithm,
IV counting frequent itemsets in a stream, clustering techniques: hierarchical, K-means, 08
clustering high dimensional data, CLIQUE and ProCLUS, frequent pattern based clustering
methods, clustering in non-euclidean space, clustering for streams and parallelism.
Frame Works and Visualization: MapReduce, Hadoop, Pig, Hive, HBase, MapR,
Sharding, NoSQL Databases, S3, Hadoop Distributed File Systems, Visualization: visual
V data analysis techniques, interaction techniques, systems and applications.
Introduction to R - R graphical user interfaces, data import and export, attribute and data 08
types, descriptive statistics, exploratory data analysis, visualization before analysis,
analytics for unstructured data.
Text books and References:
1. Michael Berthold, David J. Hand, Intelligent Data Analysis, Springer
2. Anand Rajaraman and Jeffrey David Ullman, Mining of Massive Datasets, Cambridge University Press.
3. John Garrett,Data Analytics for IT Networks : Developing Innovative Use Cases, Pearson Education
Curriculum & Evaluation Scheme IT & CSI (V & VI semester) 23
Unit-IV: Frequent Itemsets and
Clustering
1 Mining Frequent Itemsets
Frequent itemset mining is a popular data mining task that involves identifying sets of items that frequently
co-occur in a given dataset. In other words, it involves finding the items that occur together frequently and
then grouping them into sets of items. One way to approach this problem is by using the Apriori algorithm,
which is one of the most widely used algorithms for frequent itemset mining.
S
The Apriori algorithm works by iteratively generating candidate itemsets and then checking their fre-
.R
quency against a minimum support threshold. The algorithm starts by generating all possible itemsets of
size 1 and counting their frequencies in the dataset. The itemsets that meet the minimum support threshold
are then selected as frequent itemsets. The algorithm then proceeds to generate candidate itemsets of size
Dr
2 from the frequent itemsets of size 1 and counts their frequencies. This process is repeated until no more
to the potentially large number of candidate itemsets that need to be generated and counted. Point-wise
20
frequent itemset mining is a more efficient alternative that can reduce the computational complexity of the
Point-wise frequent itemset mining works by iterating over the transactions in the dataset and identifying
12
the itemsets that occur in each transaction. For each transaction, the algorithm generates a bitmap vector
where each bit corresponds to an item in the dataset, and its value is set to 1 if the item occurs in the
ril
transaction and 0 otherwise. The algorithm then performs a bitwise AND operation between the bitmap
Ap
vectors of each transaction to identify the itemsets that occur in all the transactions. The itemsets that meet
The advantage of point-wise frequent itemset mining is that it avoids generating candidate itemsets that
are not present in the dataset, thereby reducing the number of itemsets that need to be generated and
counted. Additionally, point-wise frequent itemset mining can be parallelized, making it suitable for mining
In summary, point-wise frequent itemset mining is an efficient alternative to the Apriori algorithm for
3
frequent itemset mining. It works by iterating over the transactions in the dataset and identifying the
itemsets that occur in each transaction, thereby avoiding the generation of candidate itemsets that are not
Market-based modeling is a technique used in economics and business to analyze and simulate the behavior of
markets, particularly in relation to the supply and demand of goods and services. This modeling technique
involves creating mathematical models that can simulate how different market participants (consumers,
S
producers, and intermediaries) interact with each other in a market setting.
One of the most common market-based models is the supply and demand model, which assumes that the
.R
price of a good or service is determined by the balance between its supply and demand. In this model, the
price of a good or service will rise if the demand for it exceeds its supply, and will fall if the supply exceeds
the demand. Dr
Another popular market-based model is the game theory model, which is used to analyze how different
4/
participants in a market interact with each other. Game theory models assume that market participants are
rational and act in their own self-interest, and seek to identify the strategies that each participant is likely
02
Market-based models can be used to analyze a wide range of economic phenomena, from the pricing
of individual goods and services to the behavior of entire industries and markets. They can also be used
12
to test the potential impact of various policies and interventions on the behavior of markets and market
participants.
ril
Overall, market-based modeling is a powerful tool for understanding and predicting the behavior of
markets and the economy as a whole. By creating mathematical models that simulate the behavior of
Ap
market participants and the interactions between them, economists and business analysts can gain valuable
insights into the workings of markets, and develop strategies for managing and optimizing their performance.
3 Apriori Algorithm
The Apriori algorithm is a popular algorithm used in data mining and machine learning to discover frequent
itemsets in large transactional datasets. It was proposed by Agrawal and Srikant in 1994 and is widely used
4
in association rule mining, market basket analysis, and other data mining applications.
The Apriori algorithm uses a bottom-up approach to generate all frequent itemsets by first identifying
frequent individual items and then using those items to generate larger itemsets. The algorithm works by
First, the algorithm scans the entire dataset to identify all individual items and their frequency of
occurrence. This information is used to generate the initial set of frequent itemsets.
Next, the algorithm uses a level-wise search strategy to generate larger itemsets by combining fre-
quent itemsets from the previous level. The algorithm starts with two-itemsets and then progressively
S
generates larger itemsets until no more frequent itemsets can be found.
At each level, the algorithm prunes the search space by eliminating itemsets that cannot be frequent
.R
based on the minimum support threshold. This is done using the Apriori principle, which states that
Once all frequent itemsets have been identified, the Apriori algorithm can be used to generate association
2
rules that describe the relationships between different items in the dataset. An association rule is a statement
20
of the form X − > Y, where X and Y are itemsets and X is a subset of Y. The rule indicates that there is a
The strength of an association rule is measured using two metrics: support and confidence. Support is
the percentage of transactions in the dataset that contain both X and Y, while confidence is the percentage
Overall, the Apriori algorithm is a powerful tool for discovering frequent itemsets and association rules
Ap
in large datasets. By identifying patterns and relationships between different items in the dataset, it can
be used to gain valuable insights into consumer behavior, market trends, and other important business and
economic phenomena.
Handling large datasets in main memory can be a challenging task, as the amount of memory available on
most computer systems is often limited. However, there are several techniques and strategies that can be
5
used to effectively manage and analyze large datasets in main memory:
Use data compression: Data compression techniques can be used to reduce the amount of memory
required to store a dataset. Techniques such as gzip or bzip2 can compress text data, while binary
Use data partitioning: Large datasets can be partitioned into smaller, more manageable subsets,
which can be processed and analyzed in main memory. This can be done using techniques such as
Use data sampling: Data sampling can be used to select a representative subset of data for analysis,
S
without requiring the entire dataset to be loaded into memory. Random sampling, stratified sampling,
.R
and cluster sampling are some of the commonly used sampling techniques.
Use in-memory databases: In-memory databases can be used to store large datasets in main
Dr
memory for faster querying and analysis. Examples of in-memory databases include Apache Ignite,
large datasets across multiple processors or cores. This can be done using libraries like Apache Spark,
20
Use data streaming: Data streaming techniques can be used to process large datasets in real-time
,
12
by processing data as it is generated, rather than storing it in memory. Apache Kafka, Apache Flink,
and Apache Storm are some of the popular data streaming platforms.
ril
Overall, effective management of large datasets in main memory requires a combination of data compres-
Ap
sion, partitioning, sampling, in-memory databases, parallel processing, and data streaming techniques. By
leveraging these techniques, it is possible to effectively analyze and process large datasets in main memory,
A limited pass algorithm is a technique used in data processing and analysis to efficiently process large
6
In a limited pass algorithm, the dataset is processed in a fixed number of passes or iterations, where each
pass involves processing a subset of the data. The algorithm ensures that each pass is designed to capture
the relevant information needed for the analysis, while minimizing the memory required to store the data.
For example, a limited pass algorithm for processing a large text file could involve reading the file in chunks
or sections, processing each section in memory, and then discarding the processed data before moving onto
the next section. This approach enables the algorithm to handle large datasets that cannot be loaded entirely
into memory.
Limited pass algorithms are often used in situations where the data cannot be stored in main memory,
or when the processing of the data requires significant computational resources. Examples of applications
S
that use limited pass algorithms include text processing, machine learning, and data mining.
While limited pass algorithms can be useful for processing large datasets with limited memory resources,
.R
they can also be less efficient than algorithms that can process the entire dataset in a single pass. Therefore,
it is important to carefully design the algorithm to ensure that it can capture the relevant information needed
Dr
for the analysis, while minimizing the number of passes required to process the data.
4/
6 Counting Frequent Itemsets in a Stream
2
Counting frequent itemsets in a stream is a problem of finding the most frequent itemsets in a continuous
20
stream of transactions. This problem is commonly known as the Frequent Itemset Mining problem. Here
1. Initialize a hash table to store the counts of each itemset. The size of the hash table should be limited
3. Generate all the possible itemsets from the transaction. This can be done using the Apriori algorithm,
5. Prune infrequent itemsets from the hash table. An itemset is infrequent if its count is less than a
predefined threshold.
7
7. Output the frequent itemsets that remain in the hash table after processing all the transactions.
The main challenge in counting frequent itemsets in a stream is to keep track of the changing frequencies
of the itemsets as new transactions arrive. This can be done efficiently using the hash table to store the
counts of the itemsets. However, the hash table can become too large if the number of distinct itemsets is
too large. To prevent this, the hash table can be limited in size by using a hash function that maps each
itemset to a fixed number of hash buckets. The size of the hash table can be adjusted dynamically based on
Another challenge in counting frequent itemsets in a stream is to choose the threshold for the minimum
count of an itemset to be considered frequent. The threshold should be set high enough to exclude infrequent
S
itemsets, but low enough to include all the important frequent itemsets. The threshold can be determined
.R
using heuristics or by using machine learning techniques to learn the optimal threshold from the data.
7 Clustering Techniques Dr
Clustering techniques are used to group similar data points together in a dataset based on their similarity
4/
or distance measures. Here are some popular clustering techniques:
2
20
This is a popular clustering algorithm that partitions a dataset into K clusters based on the mean dis-
,
12
tance of the data points to their assigned cluster centers. It involves an iterative process of assigning data
points to clusters and updating the cluster centers until convergence. K-Means is commonly used in image
K-Means clustering is a popular unsupervised machine learning algorithm that partitions a dataset into k
Assign each data point to the nearest cluster centroid based on its distance.
Calculate the new cluster centroids based on the mean of all data points assigned to that cluster.
8
Repeat steps 2-3 until the cluster centroids no longer change significantly, or a maximum number of
iterations is reached.
The distance metric used for step 2 is typically the Euclidean distance, but other distance metrics can
be used as well.
The K-Means algorithm aims to minimize the sum of squared distances between each data point and
its assigned cluster centroid. This objective function is known as the within-cluster sum of squares
To determine the optimal number of clusters, a common approach is to use the elbow method. This
S
involves plotting the WCSS or SSE against the number of clusters and selecting the number of clusters
.R
at the ”elbow” point, where the rate of decrease in WCSS or SSE begins to level off.
K-Means is a computationally efficient algorithm that can scale to large datasets. It is particularly useful
Dr
when the data is high-dimensional and traditional clustering algorithms may be too slow. However, K-Means
requires the number of clusters to be pre-defined and may converge to a suboptimal solution if the initial
4/
cluster centroids are not well chosen. It is also sensitive to non-linear data and may not work well with such
9
Advantages: Disadvantages:
easy to understand and implement, making it Means requires the number of clusters to be
a popular choice for clustering tasks. pre-defined, which can be a challenge when the
datasets. It is particularly useful when the data Sensitive to initial cluster centers: K-Means is
is high-dimensional and traditional clustering sensitive to the initial placement of cluster cen-
S
algorithms may be too slow. ters and can converge to a suboptimal solution
.R
K-Means works well with circular or spherical Can converge to a local minimum: K-Means
clusters, making it suitable for datasets that ex- can converge to a local minimum rather than
clustering solution.
4/
Provides a clear and interpretable result: K-
Means provides a clear and interpretable clus- Not suitable for non-linear data: K-Means as-
2
tering result, where each data point is assigned sumes that the data is linearly separable and
20
to one of the k clusters. may not work well with non-linear data.
,
12
In summary, K-Means is a simple and fast clustering algorithm that works well with circular or spherical
clusters. However, it requires the number of clusters to be pre-defined and may converge to a suboptimal
ril
solution if the initial cluster centers are not well chosen. It is also sensitive to non-linear data and may not
This technique builds a hierarchy of clusters by recursively dividing or merging clusters based on their
data point starts in its own cluster, and then pairs of clusters are successively merged until all data points
belong to a single cluster. Divisive clustering starts with all data points in a single cluster and recursively
divides them into smaller clusters. Hierarchical clustering is useful in gene expression analysis, social network
10
analysis, and image analysis.
This technique identifies clusters based on the density of data points. It assumes that clusters are areas of
higher density separated by areas of lower density. Density-based clustering algorithms, such as DBSCAN
(Density-Based Spatial Clustering of Applications with Noise), group together data points that are closely
packed together and separate outliers. Density-based clustering is commonly used in image processing,
S
7.4 Gaussian Mixture Models:
.R
This technique models the distribution of data points using a mixture of Gaussian probability distributions.
Each component of the mixture represents a cluster, and the algorithm estimates the parameters of the
mixture using the Expectation-Maximization algorithm. Gaussian Mixture Models are commonly used in
Dr
image segmentation, handwriting recognition, and speech recognition.
4/
7.5 Spectral Clustering:
2
This technique converts the data points into a graph and then partitions the graph into clusters based
20
on the eigenvalues and eigenvectors of the graph Laplacian matrix. Spectral clustering is useful in image
Each clustering technique has its own strengths and weaknesses, and the choice of clustering algorithm
depends on the nature of the data, the clustering objective, and the computational resources available.
ril
Clustering high-dimensional data is a challenging task because the distance or similarity measures used in
most clustering algorithms become less meaningful in high-dimensional space. Here are some techniques for
11
8.1 Dimensionality Reduction:
High-dimensional data can be transformed into a lower-dimensional space using dimensionality reduction
techniques, such as Principal Component Analysis (PCA) or t-SNE (t-distributed Stochastic Neighbor Em-
bedding). Dimensionality reduction can help to reduce the curse of dimensionality and make the clustering
Not all features in high-dimensional data are equally informative. Feature selection techniques can be used
to identify the most relevant features for clustering and discard the redundant or noisy features. This can
S
help to improve the clustering accuracy and reduce the computational cost.
.R
8.3 Subspace Clustering:
Dr
Subspace clustering is a clustering technique that identifies clusters in subspaces of the high-dimensional
space. This technique assumes that the data points lie in a union of subspaces, each of which represents
4/
a cluster. Subspace clustering algorithms, such as CLIQUE (CLustering In QUEst), identify the subspaces
Density-based clustering algorithms, such as DBSCAN, can be used for clustering high-dimensional data by
,
12
defining the density of data points in each dimension. The clustering algorithm identifies regions of high
Ensemble clustering combines multiple clustering algorithms or different parameter settings of the same
algorithm to improve the clustering performance. Ensemble clustering can help to reduce the sensitivity of
Deep learning-based clustering techniques, such as Deep Embedded Clustering (DEC) and Autoencoder-
based Clustering (AE-Clustering), use neural networks to learn a low-dimensional representation of high-
12
dimensional data and cluster the data in the reduced space. These techniques have shown promising results in
clustering high-dimensional data in various domains, including image analysis and gene expression analysis.
Clustering high-dimensional data requires careful consideration of the choice of clustering algorithm,
feature selection or dimensionality reduction technique, and parameter settings. A combination of different
CLIQUE (CLustering In QUEst) and ProCLUS are two popular subspace clustering algorithms for high-
dimensional data.
S
CLIQUE is a density-based algorithm that works by identifying dense subspaces in the data. It assumes
.R
that clusters exist in subspaces of the data that are dense in at least k dimensions, where k is a user-defined
parameter. The algorithm identifies all possible dense subspaces by enumerating all combinations of k
dimensions and checking if the corresponding subspaces are dense. It then merges the overlapping subspaces
Dr
to form clusters. CLIQUE is efficient for high-dimensional data because it only considers a small number of
dimensions at a time.
4/
ProCLUS (PROjective CLUSters) is a subspace clustering algorithm that works by identifying clusters
2
in a low-dimensional projection of the data. It first selects a random projection matrix and projects the data
20
onto a lower-dimensional space. It then uses K-Means clustering to cluster the projected data. The algorithm
iteratively refines the projection matrix and re-clusters the data until convergence. The final clusters are
,
projected back to the original high-dimensional space. ProCLUS is effective for high-dimensional data
12
because it reduces the dimensionality of the data while preserving the clustering structure.
Both CLIQUE and ProCLUS are designed to handle high-dimensional data by identifying clusters in
ril
subspaces of the data. They are effective for clustering data that have a natural subspace structure. However,
Ap
they may not work well for data that do not have a clear subspace structure or when the data points are
widely spread out in the high-dimensional space. It is important to carefully choose the appropriate algorithm
Frequent pattern-based clustering methods combine frequent pattern mining with clustering techniques to
identify clusters based on frequent patterns in the data. Here are some examples of frequent pattern-based
13
clustering methods:
1. Frequent Pattern-based Clustering: is a clustering algorithm that uses frequent pattern mining to
identify clusters in transactional data. The algorithm first identifies frequent itemsets in the data
using Apriori or FP-Growth algorithms. It then constructs a graph where each frequent itemset is a
node, and the edges represent the overlap between the itemsets. The graph is partitioned into clusters
using a graph clustering algorithm. The resulting clusters are then used to assign objects to clusters
2. Frequent Pattern-based Clustering Method: is a clustering algorithm that uses frequent pattern mining
S
to identify clusters in high-dimensional data. The algorithm first discretizes the continuous data into
categorical data. It then uses Apriori or FP-Growth algorithms to identify frequent itemsets in the
.R
categorical data. The frequent itemsets are used to construct a binary matrix that represents the
membership of objects in the frequent itemsets. The binary matrix is clustered using a standard
Dr
clustering algorithm, such as K-Means or Hierarchical clustering. The resulting clusters are then used
pattern mining with pattern combination techniques to identify clusters in transactional data. The
20
algorithm first identifies frequent itemsets in the data using Apriori or FP-Growth algorithms. It
then uses pattern combination techniques, such as Minimum Description Length (MDL) or Bayesian
,
Information Criterion (BIC), to generate composite patterns from the frequent itemsets. The composite
12
patterns are then used to construct a graph, which is partitioned into clusters using a graph clustering
algorithm.
ril
Frequent pattern-based clustering methods are effective for identifying clusters based on frequent patterns
Ap
in the data. They can be applied to a wide range of data types, including transactional data and high-
dimensional data. However, these methods may suffer from the curse of dimensionality when applied to
high-dimensional data. It is important to carefully select the appropriate frequent pattern mining and
clustering techniques based on the characteristics of the data and the clustering objectives.
14
10 Clustering in non-Euclidean space
Clustering in non-Euclidean space refers to the clustering of data points that are not represented in the
Euclidean space, such as graphs, time series, or text data. Traditional clustering algorithms, such as K-
Means and Hierarchical clustering, assume that the data points are represented in the Euclidean space and
use distance metrics, such as Euclidean distance or cosine similarity, to measure the similarity between data
points. However, in non-Euclidean spaces, the notion of distance is different, and distance-based clustering
S
1. Spectral clustering: Spectral clustering is a popular clustering algorithm that can be applied to data
represented in non-Euclidean spaces, such as graphs or time series. It uses the eigenvalues and eigen-
.R
vectors of the Laplacian matrix of the data to identify clusters. Spectral clustering converts the data
points into a graph representation and then computes the Laplacian matrix of the graph. The eigen-
Dr
vectors of the Laplacian matrix are used to embed the data points into a lower-dimensional space,
where clustering is performed using a standard clustering algorithm, such as K-Means or Hierarchical
4/
clustering.
2
that can be applied to data represented in non-Euclidean spaces. It does not rely on a distance
metric and can cluster data points based on their density. DBSCAN identifies clusters by defining two
,
12
parameters: the minimum number of points required to form a cluster and a radius that determines
the neighborhood of a point. DBSCAN labels each point as either a core point, a border point, or a
noise point, based on its neighborhood. The core points are used to form clusters.
ril
3. Topic modeling: Topic modeling is a clustering method that can be applied to text data, which is
Ap
typically represented in a non-Euclidean space. Topic modeling identifies latent topics in the text data
by analyzing the co-occurrence of words. It represents each document as a distribution over topics,
and each topic as a distribution over words. The resulting topic distribution of each document can be
Clustering in non-Euclidean spaces requires careful consideration of the appropriate algorithms and tech-
niques that are suitable for the specific data type. Spectral clustering and DBSCAN are effective for clustering
15
data represented as graphs or time series, while topic modeling is suitable for text data. Other approaches,
such as manifold learning and kernel methods, can also be used for clustering in non-Euclidean spaces.
Clustering for streams and parallelism are two important considerations for clustering large datasets. Stream
data refers to data that arrives continuously and in real-time, while parallelism refers to the ability to
S
1. Online clustering: Online clustering is a technique that can be applied to streaming data. It updates
.R
the clustering model continuously as new data arrives. Online clustering algorithms, such as BIRCH
and CluStream, are designed to handle data streams and can scale to large datasets. These algo-
rithms incrementally update the cluster model as new data arrives and discard outdated data points
Dr
to maintain the cluster model’s accuracy and efficiency.
4/
2. Parallel clustering: Parallel clustering refers to the use of multiple computing resources, such as multiple
processors or computing clusters, to speed up the clustering process. Parallel clustering algorithms,
2
such as K-Means Parallel, Hierarchical Parallel, and DBSCAN Parallel, distribute the clustering task
20
across multiple computing resources. These algorithms partition the data into smaller subsets and
assign each subset to a separate computing resource. The resulting clusters are then merged to produce
,
12
3. Distributed clustering: Distributed clustering refers to the use of multiple computing resources that
ril
are distributed across different physical locations, such as different data centers or cloud resources.
Ap
Distributed clustering algorithms, such as MapReduce and Hadoop, distribute the clustering task
across multiple computing resources and handle data that is too large to fit into a single computing
resource’s memory. These algorithms partition the data into smaller subsets and assign each subset to
a separate computing resource. The resulting clusters are then merged to produce the final clustering
result.
Clustering for streams and parallelism requires careful consideration of the appropriate algorithms and
techniques that are suitable for the specific clustering objectives and data types. Online clustering is effective
16
for clustering streaming data, while parallel clustering and distributed clustering can speed up the clustering
Q1: Write R function to check whether the given number is prime or not.
flag = 0
S
# check for factors flag = 1
for(i in 2:(num-1)) {
.R
if ((num %% i) == 0)
flag = 0
break
}
Dr
4/
}
2
}
20
if(num == 2) flag = 1
if(flag == 1)
,
else
17
Apriori algorithm:—The apriori algorithm solves the frequent item sets problem. The algorithm ana-
lyzes a data set to determine which combinations of items occur together frequently. The Apriori algorithm
is at the core of various algorithms for data mining problems. The best known problem is finding the asso-
Numerical:
Given:
S
ITERATION:1
STEP 1: (C1)
ITERATION 2:
.R
Itemsets Counts STEP 3: (C2)
A 1 Itemsets Counts
C 2 STEP 2: (L2) STEP 4: (L2)
E, K 4
D 1 ITERATION 3:
E
K
I
4
1
5
Itemsets Counts
E
K
4
5
E, M
E, O
E, Y
Dr
2
3
2
Itemsets Counts
E, K
E, O
4
3
STEP 5: (C3)
Itemsets Counts
E, K, O 3
M 3 K, M 3 K, M 3
4/
M 3 K, M, O 1
O 3 K, O 3 K, O 3
N 2 K, M, Y 2
Y 3 K, Y 3 K, Y 3
O 3
2
M, O 1
U 1
M, Y 2
20
Y 3
O, Y 2
STEP 6: (L3)
Itemsets Counts
,
E, K, O 3
12
ASSOCIATION RULE:
Ap
18
5. K → [E, O] = 3/5 = 60%
S
.R
Dr
2 4/
, 20
12
ril
Ap
19
Printed Page: 1 of 2
Subject Code: KIT601
0Roll No: 0 0 0 0 0 0 0 0 0 0 0 0 0
BTECH
(SEM VI) THEORY EXAMINATION 2021-22
DATA ANALYTICS
SECTION A
S
(d) What is multivariate analysis? 2
(e) Give the full form of RTAP and discuss its application. 3
.R
(f) What is the role of sampling data in a stream? 3
(g) Discuss the use of limited pass algorithm. 4
(h) What is the principle behind hierarchical clustering technique? 4
(i)
(j)
Dr
List five R functions used in descriptive statistics.
List the names of any 2 visualization tools.
5
5
4/
SECTION B
2
Qno Questions CO
(a) Explain the process model and computation model for Big data 1
,
platform.
12
SECTION C
3. Attempt any one part of the following: 10*1 = 10
Qno Questions CO
(a) Explain the various phases of data analytics life cycle. 1
(b) Explain modern data analytics tools in detail. 1
BTECH
(SEM VI) THEORY EXAMINATION 2021-22
DATA ANALYTICS
S
(b) A database has 5 transactions. Let min_sup=60% and min_conf=80%. 4
.R
TID Items_Bought
T100 {M, O, N, K, E, Y}
T200 {D, O, N, K, E, Y}
T300 Dr {M, A, K, E}
T400 {M, U, C, K, Y}
T500 {C, O, O, K, I, E}
2 4/
ii) List all the strong association rules (with support s and confidence
c).
,
12
S
.R
Dr
2 4/
, 20
12
ril
Ap
24
DATABASE
SYSTEMS
What is Frequent Itemset Mining?
GROUP
S
• Task 2: find all rules that correlate the presence of one set of items with
that of another set of items in the transaction database.
.R
– E.g.: 98% of people buying tires and auto accessories also get automotive service
done
• Applications: Basket data analysis, cross-marketing, catalog design,
Dr
loss-leader analysis, clustering, classification, recommendation systems,
etc.
4/
Frequent Itemset Mining Introduction 3
2
20
DATABASE
SYSTEMS
Example: Basket Data Analysis
,
GROUP
12
• Transaction database
D= {{butter, bread, milk, sugar};
ril
{eggs};
{butter, flour, milk, salt, sugar}}
1) Introduction
– Transaction databases, market basket data analysis
2) Mining Frequent Itemsets
– Apriori algorithm, hash trees, FP-tree
3) Simple Association Rules
– Basic notions, rule generation, interestingness measures
4) Further Topics
– Hierarchical Association Rules
• Motivation, notions, algorithms, interestingness
S
– Quantitative Association Rules
• Motivation, basic idea, partitioning numerical attributes, adaptation of
.R
apriori algorithm, interestingness
5) Extensions and Summary
Dr
4/
Outline 5
2
Notions
DATABASE
SYSTEMS
,
GROUP
12
𝐼
• Transaction 𝑇 contains an itemset 𝑋: 𝑋 ⊆ 𝑇
• The items in transactions and itemsets are sorted lexicographically:
– itemset 𝑋 = (𝑥1, 𝑥2, … , 𝑥𝑘 ), where 𝑥1 𝑥2 … 𝑥𝑘
• Length of an itemset: number of elements in the itemset
• k-itemset: itemset of length k
• The support of an itemset X is defined as: 𝑠𝑢𝑝𝑝𝑜𝑟𝑡 𝑋 = 𝑇 ∈ 𝐷|𝑋 ⊆ 𝑇
• Frequent itemset: an itemset X is called frequent for database 𝐷 iff it is
contained in more than 𝑚𝑖𝑛𝑆𝑢𝑝 many transactions: 𝑠𝑢𝑝𝑝𝑜𝑟𝑡(𝑋) ≥
𝑚𝑖𝑛𝑆𝑢𝑝
• Naïve Algorithm
– count the frequency of all possible subsets of 𝐼 in the database
too expensive since there are 2m such itemsets for |𝐼| = 𝑚 items
cardinality of power set
• The Apriori principle (anti-monotonicity):
Any non-empty subset of a frequent itemset is frequent, too!
A ⊆ I with support A ≥ minSup ⇒ ∀A′ ⊂ A ∧ A′ ≠ ∅: support A′ ≥ minSup
Any superset of a non-frequent itemset is non-frequent, too!
A ⊆ I with support A < minSup ⇒ ∀A′ ⊃ A: support A′ < minSup
ABCD not frequent
S
• Method based on the apriori principle ABC ABD ACD BCD
– First count the 1-itemsets, then the 2-itemsets,
.R
AB AC AD BC BD CD
then the 3-itemsets, and so on
A B C D
– When counting (k+1)-itemsets, only consider those
(k+1)-itemsets where all subsets of length k have been
Dr Ø
determined as frequent in the previous step
4/
Frequent Itemset Mining Algorithms Apriori Algorithm 7
2
20
DATABASE
SYSTEMS
The Apriori Algorithm
,
GROUP
12
L1 = {frequent items}
for (k = 1; Lk !=; k++) do begin
// JOIN STEP: join Lk with itself to produce Ck+1
produce // PRUNE STEP: discard (k+1)-itemsets from Ck+1 that
candidates contain non-frequent k-itemsets as subsets
Ck+1 = candidates generated from Lk
S
p Lk=3 (A, C, F)
insert into Ck+1
.R
select p.i1, p.i2, …, p.ik–1, p.ik, q.ik (A, C, F, G) Ck+1=4
from Lk : p, Lk : q Dr q Lk=3 (A, C, G)
where p.i1=q.i1, …, p.ik –1 =q.ik–1, p.ik < q.ik
4/
Frequent Itemset Mining Algorithms Apriori Algorithm 9
2
20
DATABASE
SYSTEMS
Generating Candidates (Prune Step)
,
GROUP
12
– Naïve: Check support of every itemset in 𝐶𝑘+1 inefficient for huge 𝐶𝑘+1
– Instead, apply Apriori principle first: Remove candidate (k+1) -itemsets
Ap
S
{3 5} {3 5} {3 5} 2
{3 6} {3 6} {3 6} 1 𝐿2 ⋈ 𝐿2
.R
{5 6} {5 6} {5 6} 1
C3 itemset C3 itemsetC3 itemset count L3 itemset count
{1 3 5} {1 3 5} {1 3 5} 1 {2 3 5} 2
{1 3 6} prune C2 {1 3 6} ✗ scan D {2 3 5} 2
{1 5 6} {1 5 6} ✗
Dr 𝐿3 ⋈ 𝐿3
{2 3 5} {2 3 5}
4/
C4 is empty
Frequent Itemset Mining Algorithms Apriori Algorithm 11
2
Candidates?
DATABASE
SYSTEMS
,
GROUP
12
• Method: Hash-Tree
– Candidate itemsets are stored in a hash-tree
– Leaf nodes of hash-tree contain lists of itemsets and their support (i.e.,
counts)
– Interior nodes contain hash tables
– Subset function finds all the candidates contained in a transaction
012
012 012 012
e.g. for 3-Itemsets
h(K) = K mod 3 (3 6 7) 0 1 2 (3 5 7) (7 9 12) (1 4 11) (7 8 9) (2 3 8) 0 1 2 (2 5 6)
(3 5 11) (1 6 11) (1 7 9) (1 11 12) (5 6 7) (2 5 7)
(5 8 11)
(3 4 15) (3 7 11) (2 4 6) (2 4 7)
(3 4 11) (2 7 9) (5 7 10)
(3 4 8)
S
012 012 012
.R
for 3-Itemsets
(3 6 7) 0 1 2 (3 5 7) (7 9 12) (1 4 11) (7 8 9) (2 3 8) 0 1 2 (2 5 6)
h(K) = K mod 3 (3 5 11) (1 6 11) (1 7 9) (1 11 12) (5 6 7) (2 5 7)
(5 8 11)
(3 4 15) (3 7 11)
(3 4 11)
(3 4 8)
Dr (2 4 6) (2 4 7)
(2 7 9) (5 7 10)
4/
Frequent Itemset Mining Algorithms Apriori Algorithm 13
2
20
DATABASE
SYSTEMS
Hash-Tree – Counting
,
GROUP
12
• Search all candidate itemsets contained in a transaction T = (t1 t2 ... tn) for a
current itemset length of k
ril
• At the root
Ap
S
• Needs n or n+1 scans, n is the length of the longest pattern
.R
Is it possible to mine the complete set of frequent itemsets without
candidate generation?
Dr
4/
Frequent Itemset Mining Algorithms Apriori Algorithm 15
2
Candidate Generation
DATABASE
SYSTEMS
,
GROUP
12
• Idea:
– Compress database into FP-tree, retaining the itemset association
information
– Divide the compressed database into conditional databases, each associated
with one frequent item and mine each such database separately.
S
header table:
1&2 item frequency
.R
f 4
c 4 sort items in the order
a 3 of descending support
minSup=0.5 b 3 Dr
m 3
p 3
4/
Frequent Itemset Mining Algorithms FP-Tree 17
2
DB
DATABASE
SYSTEMS
,
GROUP
12
3. Scan DB again, construct FP-tree starting with most frequent item per transaction
TID items bought (ordered) frequent
items for each transaction only
100 {f, a, c, d, g, i, m, p} {f, c, a, m, p} keep its frequent items
200 {a, b, c, f, l, m, o} {f, c, a, b, m} sorted in descending
300 {b, f, h, j, o} {f, b} order of their frequencies
400 {b, c, k, s, p} {c, b, p}
500 {a, f, c, e, l, p, m, n} {f, c, a, m, p}
header table:
item frequency 3a
1&2 f 4
for each transaction build a path in the FP-tree:
c 4
a 3 - If a path with common prefix exists:
b 3 increment frequency of nodes on this path
m 3 and append suffix
p 3 - Otherwise: create a new branch
Frequent Itemset Mining Algorithms FP-Tree 18
Construct FP-tree from a Transaction
DB
DATABASE
SYSTEMS
GROUP
S
header table: 3a
1&2
item frequency head
.R
a:3 p:1
f 4
header table c 4
references the a 3 m:2 b:1
occurrences of the
frequent items in the
b
m
3
3
Dr p:2 m:1
FP-tree
4/
p 3
Frequent Itemset Mining Algorithms FP-Tree 19
2
20
DATABASE
SYSTEMS
Benefits of the FP-tree Structure
,
GROUP
12
• Completeness:
ril
• Compactness
– reduce irrelevant information—infrequent items are gone
– frequency descending ordering: more frequent items are more likely to be
shared
– never be larger than the original database (if not count node-links and
counts)
– Experiments demonstrate compression ratios over 100
S
.R
Dr
4/
Frequent Itemset Mining Algorithms FP-Tree 21
2
20
DATABASE
SYSTEMS
Major Steps to Mine FP-tree
,
GROUP
12
obtained so far
– If the conditional FP-tree contains a single path, simply enumerate all the
patterns
S
f 4 item cond. pattern base
c 4 c:3 b:1 b:1 f {}
.R
a 3 c f:3, {}
b 3 a:3 p:1 a fc:3
m 3 b fca:1, f:1, c:1
p 3
m:2 b:1 Dr m
p
fca:2, fcab:1
fcam:2, cb:1
p:2 m:1
4/
Frequent Itemset Mining Algorithms FP-Tree 23
2
Pattern Bases
DATABASE
SYSTEMS
,
GROUP
12
• Node-link property
ril
– For any frequent item ai, all the possible frequent patterns that contain ai
can be obtained by following ai's node-links, starting from ai's head in the
Ap
FP-tree header
• Prefix path property
– To calculate the frequent patterns for a node ai in a path P, only the prefix
sub-path of ai in P needs to be accumulated, and its frequency count should
carry the same count as node ai.
S
item cond. pattern base item frequency {}|m
f {} f 3 ..
.R
c f:3 c 3 ..
a fc:3 f:3
a 3 ..
b fca:1, f:1, c:1 b 1✗
m
p
fca:2, fcab:1
fcam:2, cb:1
Dr c:3
a:3
4/
Frequent Itemset Mining Algorithms FP-Tree 25
2
Conditional FP-tree
DATABASE
SYSTEMS
,
GROUP
12
c:3 c:3
a:3
Frequent Itemset Mining Algorithms FP-Tree 26
DATABASE
SYSTEMS
Major Steps to Mine FP-tree
GROUP
example:
m-conditional FP-tree All frequent patterns
{}|m concerning m
S
m,
f:3 just a single path
fm, cm, am,
.R
c:3 fcm, fam, cam,
fcam
a:3 Dr
4/
Frequent Itemset Mining Algorithms FP-Tree 27
2
20
DATABASE
SYSTEMS
FP-tree: Full Example
,
GROUP
12
database:
TID items bought (ordered) frequent items
ril
c:2
conditional pattern base:
item cond. pattern base
f {}
b f:2, {}
c fb:2, b:1
c:2
S
f:2 b:2
.R
{{f}}
{{b},{fb}}
Dr {}|fc = {}
{{fc}}
{}|bc
{{bc},{fbc}}
f:2
4/
Frequent Itemset Mining Algorithms FP-Tree 29
2
Growth
DATABASE
SYSTEMS
,
GROUP
12
in B.
70
tree-projection
Run time(sec.)
60
50 D1 FP-grow th runtime
40 D1 Apriori runtime
30
20
10
0
• Reasoning
0 0,5 1 1,5 2 2,5 3
Support threshold(%)
S
– No candidate generation, no candidate test
.R
• Apriori algorithm has to proceed breadth-first
– Use compact data structure
– Eliminate repeated database scan
– Basic operation is counting and FP-tree building
Dr
4/
Frequent Itemset Mining Algorithms FP-Tree 31
2
20
DATABASE
SYSTEMS
Maximal or Closed Frequent Itemsets
,
GROUP
12
1) Introduction
– Transaction databases, market basket data analysis
2) Mining Frequent Itemsets
– Apriori algorithm, hash trees, FP-tree
3) Simple Association Rules
– Basic notions, rule generation, interestingness measures
4) Further Topics
– Hierarchical Association Rules
• Motivation, notions, algorithms, interestingness
S
– Quantitative Association Rules
• Motivation, basic idea, partitioning numerical attributes, adaptation of
.R
apriori algorithm, interestingness
5) Extensions and Summary
Dr
4/
Outline 33
2
Introduction
DATABASE
SYSTEMS
,
GROUP
12
• Transaction database:
ril
S
.R
• Association rule: An association rule is an implication of the form 𝑋 ⇒ 𝑌
where 𝑋, 𝑌 ⊆ 𝐼 are two itemsets with 𝑋 ∩ 𝑌 = ∅.
Dr
• Note: simply enumerating all possible association rules is not reasonable!
What are the interesting association rules w.r.t. 𝐷?
4/
Frequent Itemset Mining Simple Association Rules 35
2
20
DATABASE
SYSTEMS
Interestingness of Association Rules
,
GROUP
12
transaction database D:
Ap
S
• Example
frequent itemsets
.R
1-itemset count 2-itemset count 3-itemset count
{A} 3 {A, B} 3 {A, B, C} 2
{B} 4 {A, C} 2
{C} 5 {B, C} 4
Dr
rule candidates: A ⇒ 𝐵; 𝐵 ⇒ 𝐴; A ⇒ 𝐶; 𝐶 ⇒ A; 𝐵 ⇒ 𝐶; C ⇒ 𝐵;
𝐴, 𝐵 ⇒ 𝐶; 𝐴, 𝐶 ⇒ 𝐵; 𝐶, 𝐵 ⇒ 𝐴; 𝐴 ⇒ 𝐵, 𝐶; 𝐵 ⇒ 𝐴, 𝐶; 𝐶 ⇒ 𝐴, 𝐵
4/
Frequent Itemset Mining Simple Association Rules 37
2
Itemsets
DATABASE
SYSTEMS
,
GROUP
12
• Objective measures
– Two popular measurements:
– support and
– confidence
S
.R
Dr
4/
Frequent Itemset Mining Simple Association Rules 39
2
20
DATABASE
SYSTEMS
Criticism to Support and Confidence
,
GROUP
12
S
• 𝑐𝑜𝑟𝑟𝐴,𝐵 = 1 there is no correlation between the two items A and B
.R
• 𝑐𝑜𝑟𝑟𝐴,𝐵 < 1 the two items A and B are negatively correlated
Dr
4/
Frequent Itemset Mining Simple Association Rules 41
2
Correlation
DATABASE
SYSTEMS
,
GROUP
12
• Example 2: X 1 1 1 1 0 0 0 0
ril
Y 1 1 0 0 0 0 0 0
Z 0 1 1 1 1 1 1 1
Ap
1) Introduction
– Transaction databases, market basket data analysis
2) Mining Frequent Itemsets
– Apriori algorithm, hash trees, FP-tree
3) Simple Association Rules
– Basic notions, rule generation, interestingness measures
4) Further Topics
– Hierarchical Association Rules
• Motivation, notions, algorithms, interestingness
S
– Quantitative Association Rules
• Motivation, basic idea, partitioning numerical attributes, adaptation of
.R
apriori algorithm, interestingness
5) Extensions and Summary
Dr
4/
Outline 43
2
Motivation
DATABASE
SYSTEMS
,
GROUP
12
jackets jeans
• Examples
Jeans boots
jackets boots Support < minSup
• Characteristics
– Support(“outerwear boots”) is not necessarily equal to the sum
support(“jackets boots”) + support( “jeans boots”)
e.g. if a transaction with jackets, jeans and boots exists
S
– Support for sets of generalizations (e.g., product groups) is higher
than support for sets of individual items
.R
If the support of rule “outerwear boots” exceeds minsup, then the
support of rule “clothes boots” does, too
Dr
4/
Frequent Itemset Mining Further Topics Hierarchical Association Rules 45
2
20
DATABASE
SYSTEMS
Mining Multi-Level Associations
,
GROUP
12
milk bread
– Then find their lower-level “weaker” rules:
• 1.5% milk wheat bread [6%, 50%]. 3.5% 1.5% wheat white
S
.R
3.5% 1.5% minsup = 3 %
support = 6 % support = 4 %
+ takes the lower frequency of items in lower levels into consideration
Dr
4/
Frequent Itemset Mining Further Topics Hierarchical Association Rules 47
2
Reduced Support
DATABASE
SYSTEMS
,
GROUP
12
milk bread
– Then find their lower-level “weaker” rules:
• 1.5% milk wheat bread [6%, 50%]. 3.5% 1.5% wheat white
level-wise processing (breadth first)
3 approaches using reduced Support: Fraser Sunset Wonder
S
– Level-crossed association rules:
• 1.5 % milk Wonder wheat bread
.R
– Association rules with multiple, alternative hierarchies:
• 1.5 % milk Wonder bread
Dr
4/
Frequent Itemset Mining Further Topics Hierarchical Association Rules 49
2
Filtering
DATABASE
SYSTEMS
,
GROUP
12
• Example
Ap
Let 𝑋, 𝑋 ′ , 𝑌, 𝑌 ′ ⊆ 𝐼 be itemsets.
• An itemset 𝑋′ is an ancestor of 𝑋 iff there exist ancestors 𝑥1′ , … , 𝑥𝑘′ of
𝑥1 , … , 𝑥𝑘 ∈ 𝑋 and 𝑥𝑘+1 , … , 𝑥𝑛 with 𝑛 = 𝑋 such that
𝑋 ′ = {𝑥1′ , … , 𝑥𝑘′ , 𝑥𝑘+1 , … , 𝑥𝑛 }.
• Let 𝑋 ′ and 𝑌′ be ancestors of 𝑋 and 𝑌. Then we call the rules 𝑋′ 𝑌′,
𝑋𝑌′, and 𝑋′𝑌 ancestors of the rule X Y .
• The rule X´ Y´ is a direct ancestor of rule X Y in a set of rules if:
– Rule X´ Y‘ is an ancestor of rule X Y, and
– There is no rule X“ Y“ such that X“ Y“ is an ancestor of
S
X Y and X´ Y´ is an ancestor of X“ Y“
.R
• A hierarchical association rule X Y is called R-interesting if:
– There are no direct ancestors of X Y or
Dr
– The actual support is larger than R times the expected support or
– The actual confidence is larger than R times the expected confidence
4/
Frequent Itemset Mining Further Topics Hierarchical Association Rules 51
2
Confidence
DATABASE
SYSTEMS
,
GROUP
12
P(𝑧1 ) P 𝑧𝑗 ′
𝐸𝑍 ′ P 𝑍 = ′ × ⋯ × ′ ×P 𝑍
P(𝑧1 ) P(𝑧𝑗 )
where 𝑍 = 𝑋 ∪ 𝑌 = {𝑧1 , … , 𝑧𝑛 }, 𝑍 ′ = 𝑋 ′ ∪ 𝑌 ′ = {𝑧1′ , … , 𝑧𝑗′ , 𝑧𝑗+1 , … , 𝑧𝑛 } and
each 𝑧𝑖′ ∈ 𝑍 ′ is an ancestor of 𝑧𝑖 ∈ 𝑍
S
.R
Dr
[SA’95] R. Srikant, R. Agrawal: Mining Generalized Association Rules. In VLDB, 1995.
4/
Frequent Itemset Mining Further Topics Hierarchical Association Rules 53
2
Interestingness of Hierarchical
20
Association Rules:Example
DATABASE
SYSTEMS
,
GROUP
12
jackets 4
•
No rule support R-interesting?
1 clothes shoes 10 yes: no ancestors
2 outerwear shoes 9 yes:
Support > R *exp. support (wrt. rule 1) =
10
(1.6 ⋅ (20 ⋅ 10)) = 8
3 jackets shoes 4 Not wrt. support:
Support > R * exp. support (wrt. rule 1) = 3.2
Support < R * exp. support (wrt. rule 2) = 5.75
still need to check the confidence!
1) Introduction
– Transaction databases, market basket data analysis
2) Simple Association Rules
– Basic notions, rule generation, interestingness measures
3) Mining Frequent Itemsets
– Apriori algorithm, hash trees, FP-tree
4) Further Topics
– Hierarchical Association Rules
• Motivation, notions, algorithms, interestingness
S
– Multidimensional and Quantitative Association Rules
• Motivation, basic idea, partitioning numerical attributes, adaptation of
.R
apriori algorithm, interestingness
5) Summary
Dr
4/
Outline 55
2
Multi-Dimensional Association:
20
Concepts
DATABASE
SYSTEMS
,
GROUP
12
• Single-dimensional rules:
ril
S
– This is a dynamic discretization process that considers the distance between
.R
data points.
Dr
4/
Frequent Itemset Mining Extensions & Summary 57
2
20
DATABASE
SYSTEMS
Quantitative Association Rules
,
GROUP
12
• Example:
– Original database
• Static discretization
– Discretization of all attributes before mining the association rules
– E.g. by using a generalization hierarchy for each attribute
– Substitute numerical attribute values by ranges or intervals
• Dynamic discretization
– Discretization of the attributes during association rule mining
– Goal (e.g.): maximization of confidence
– Unification of neighboring association rules to a generalized rule
S
.R
Dr
4/
Frequent Itemset Mining Further Topics Quantitative Association Rules 59
2
20
DATABASE
SYSTEMS
Partitioning of Numerical Attributes
,
GROUP
12
• Solution
– First, partition the domain into many intervals
– Afterwards, create new intervals by merging adjacent interval
• Example:
S
.R
age(X,”30-34”) income(X,”24K - 48K”)
buys(X,”high resolution TV”)
Dr
4/
Frequent Itemset Mining Further Topics Quantitative Association Rules 61
2
20
DATABASE
SYSTEMS
Chapter 3: Frequent Itemset Mining
,
GROUP
12
1) Introduction
ril
Outline 62
12 Reference
[1] https://www.jigsawacademy.com/blogs/hr-analytics/data-analytics-lifecycle/
[2] https://statacumen.com/teach/ADA1/ADA1\_notes\_F14.pdf
[3] https://www.youtube.com/watch?v=fDRa82lxzaU
[4] https://www.investopedia.com/terms/d/data-analytics.asp
[5] http://egyankosh.ac.in/bitstream/123456789/10935/1/Unit-2.pdf
[6] http://epgp.inflibnet.ac.in/epgpdata/uploads/epgp\_content/computer\_science/16.\_d
S
ata\_analytics/03.\_evolution\_of\_analytical\_scalability/et/9280\_et\_3\_et.pdf
.R
[7] https://bhavanakhivsara.files.wordpress.com/2018/06/data-science-and-big-data-analy
-nieizv\_book.pdf
Dr
[8] https://www.researchgate.net/publication/317214679\_Sentiment\_Analysis\_for\_Effect
4/
ive\_Stock\_Market\_Prediction
[9] https://snscourseware.org/snscenew/files/1569681518.pdf
2
20
[10] http://csis.pace.edu/ctappert/cs816-19fall/books/2015DataScience\&BigDataAnalytics.
pdf
,
12
[11] https://www.youtube.com/watch?v=mccsmoh2\_3c
[12] https://mentalmodels4life.net/2015/11/18/agile-data-science-applying-kanban-in-the-a
ril
nalytics-life-cycle/
Ap
[13] https://www.sas.com/en_in/insights/big-data/what-is-big-data.html#:~:text=Big%20dat
a%20refers%20to%20data,around%20for%20a%20long%20time.
[14] https://www.javatpoint.com/big-data-characteristics
[15] Liu, S., Wang, M., Zhan, Y., & Shi, J. (2009). Daily work stress and alcohol use: Testing the cross-
level moderation effects of neuroticism and job involvement. Personnel Psychology,62(3), 575–597.
http://dx.doi.org/10.1111/j.1744-6570.2009.01149.x
55
[16] https://www.google.com/search?q=architecture+of+data+stream+model&sxsrf=APwXEdf9LJ8N
XMypRU-Sg28SH8m_pwiUDA:1679823244352&source=lnms&tbm=isch&sa=X&ved=2ahUKEwjGgY-epfn9
AhX5xTgGHRWjDmMQ_AUoAXoECAEQAw&biw=1366&bih=622#imgrc=wnFWJQ01p-w_jM
[17] Prof. Dr. Thomas Seidl, Frequent Itemset Mining, Knowledge Discovery in Databases, SS 2016.
********************
S
.R
Dr
2 4/
, 20
12
ril
Ap
56