Deep Learning Project Report
Deep Learning Project Report
A PROJECT REPORT
Submitted by
BACHELOR OF ENGINEERING
IN
COMPUTER SCIENCE & ENGINEERING
Chandigarh University
April, 2024
1|Page
BONAFIDE CETIFICATE
Certified that this project report “Real Time Human Emotion Detection” is the
Bonafide work of “Heemaal Jaglan” who carried out the project work under my
supervision.
SIGNATURE SIGNATURE
Dr. Sandeep Singh Kang [Link] Kaur
SUPERVISOR
HEAD OF THE DEPARTMENT
Assistant Professor
Computer Science and Engineering
Computer Science and Engineering
2|Page
ACKNOWLEDGEMENT
I sincerely express my gratitude to Er. Sukhvir Kaur for their invaluable
guidance and support throughout our project, " Real Time Human
Emotion Detection " Their insights have been instrumental in shaping
our research.
I also extend our thanks to Chandigarh University for providing the
necessary resources and a conducive learning environment.
“Excellence is not a destination; it is a continuous journey that never
ends”
3|Page
TABLE OF CONTENTS
S. No. CONTENT Page No.
A. Abstract 6
1. Introduction 7-11
1.4. Timeline 10
2.6. Goals/Objectives 18
4|Page
3.3. Analysis Of Features And Finalization Subject To 21
Constraints
5.1. Conclusion 30
B. References 32
List Of Figures
Fig 1.1 ……………………………………………………………… 10
Fig 1.2 ……………………………………………………………… 25
Fig 1.3 ……………………………………………………………… 26
List Of Tables
Table 1.1 …………………………………………………………… 11
5|Page
ABSTARCT
This project presents a real-time human emotion detection system that combines
computer vision and deep learning to analyze facial expressions and identify
emotional states such as happiness, sadness, anger, fear, and surprise. The system
uses OpenCV for real-time face detection and employs a pre-trained Convolutional
Neural Network (CNN) model to classify emotions accurately.
Live video input is processed frame by frame to detect facial landmarks, extract
features, and predict emotions, allowing for dynamic interaction between humans
and machines. This technology has wide-ranging applications in areas such as
healthcare monitoring, virtual assistants, smart classrooms, and customer service
enhancement. The project demonstrates the feasibility and effectiveness of real-time
emotion recognition in improving human-computer interaction.
6|Page
CHAPTER 1
INTRODUCTION
[Link]
In the modern era of artificial intelligence, machines are becoming more capable of understanding
and interacting with humans in intelligent ways. One of the key areas of this advancement is
emotion recognition, which focuses on detecting human emotions through facial expressions,
voice, body language, or physiological signals. This project specifically explores real-time
emotion detection using facial expressions and deep learning [Link] capturing live video
input and applying computer vision algorithms, the system identifies facial features and classifies
emotions using a trained neural network model. Real-time emotion detection systems aim to bridge
the emotional gap between humans and machines, allowing for more responsive, empathetic, and
intuitive human-computer interactions. This technology can be applied in various domains
including healthcare, education, entertainment, security, and customer service.
[Link]
Human emotions play a crucial role in communication and decision-making. Traditional computer
systems lack the ability to perceive and respond to the emotional state of users, which limits their
effectiveness in situations that require empathy or adaptive responses. The need for real-time
emotion detection arises from the growing demand for emotionally intelligent systems. In
healthcare, it can be used to monitor patient mood or detect signs of stress and depression. In
education, it can help tutors or e-learning platforms adapt their teaching based on student emotions.
In customer service, it enhances user experience by allowing systems to respond appropriately to
customer frustration or satisfaction. Moreover, in security and surveillance, detecting abnormal
emotional behavior can help identify potential threats.
a)Delayed Diagnosis & Misinterpretation – Traditional seizure detection relies on manual EEG
analysis, which can be time-consuming and prone to human error, leading to delayed or incorrect
diagnoses.
b)Limited Access to Neurologists – Many regions, especially rural areas, face a shortage of
trained neurologists, making timely seizure diagnosis and monitoring difficult.
c)High False Positives & Negatives – Existing seizure detection methods often produce
inaccurate results, leading to unnecessary anxiety or missed seizures.
7|Page
d)Real-Time Monitoring Challenges – Continuous EEG monitoring is costly and inconvenient,
requiring advanced solutions that can work efficiently in real-time.
e)Integration with Wearable Technology – The need for portable, AI-powered seizure detection
systems that integrate with wearables for better patient monitoring and immediate alerts.
f)Data Privacy & Security – As AI-based seizure detection relies on patient EEG data, ensuring
privacy, security, and ethical use of medical data remains a major challenge.
g)Cost & Accessibility – Many advanced seizure detection systems are expensive, making them
inaccessible to low-income patients and healthcare facilities.
h)Adaptability for Different Patients – Seizure patterns vary among individuals, requiring more
personalized AI models for effective detection and prediction.
i)Regulatory & Ethical Concerns – AI-driven medical applications require strict regulatory
approvals to ensure reliability, accuracy, and patient safety.
j)Lack of Public Awareness – Many patients and caregivers are unaware of modern seizure
detection technologies, limiting their adoption and effectiveness in epilepsy management.
Data Security: Emotion data, if not securely stored or transmitted, could be vulnerable to
breaches and misuse.
Bias in Emotion Recognition Models: Pre-trained models may show bias based on age,
gender, or ethnicity, leading to inaccurate results.
Contextual Misinterpretation: Emotions can vary depending on context; the same facial
expression might indicate different feelings in different situations.
Legal and Regulatory Gaps: There are limited laws or guidelines regulating the use of
facial emotion detection in public or private domains.
8|Page
1.3. Identification Of Task
[Link] Collection and Preprocessing: This step involves gathering CNN RNN datasets from
reliable medical sources, cleaning the data to remove noise and artifacts, and segmenting the
signals into meaningful time frames for analysis.
[Link] Extraction and Selection: Relevant signal characteristics such as facial features and
amplitude variations are identified. Techniques like wavelet transforms or Fourier analysis are
used to extract meaningful features, and the most significant ones are selected to enhance model
accuracy and efficiency.
[Link] Selection and Development: Choosing suitable deep learning architectures, such as
Convolutional Neural Networks (CNNs), Recurrent Neural Networks (RNNs), or Long Short-
Term Memory (LSTM) networks, is crucial. The model is trained on labeled seizure and non-
seizure data, and hyperparameters are optimized to improve performance.
[Link] and Validation: The dataset is split into training, validation, and testing sets.
Performance metrics like sensitivity, specificity, and F1-score are used to evaluate accuracy, and
the results are compared with existing seizure detection methods.
[Link]-Time Implementation and Optimization: The trained model is integrated into a real-time
monitoring system. Computational efficiency is optimized for deployment on mobile or wearable
devices, ensuring the system provides instant alerts upon seizure detection.
[Link] and Ethical Considerations: Ensuring patient data privacy and compliance with
healthcare regulations is essential. Ethical concerns related to AI-based medical decision-making
are addressed, and security measures are implemented to protect sensitive medical data.
[Link] and Future Enhancements: The system undergoes testing in clinical settings for
real-world validation.
9|Page
1.4. Timeline
Fig 1.1
Chapter 1 Introduction: This chapter introduces the project and describes the problem statement
discussed earlier in the report.
Chapter 2 Literature Review/Background Study: This chapter prevents review for various
research papers which help us to understand the problem in a better way. It also defines what has
been done to already solve the problem and what can be further done.
Chapter 3 Design Flow/ Process: This chapter presents the need and significance of the proposed
work based on literature review. Proposed objectives and methodology are explained. This presents
the relevance of the problem. It also represents logical and schematic plan to resolve the research
problem.
Chapter 4 Result Analysis and Validation: This chapter explains various performance
parameters used in implementation. Experimental results are shown in this chapter. It explains the
meaning of the results and why they matter.
Chapter 5 Conclusion and future scope: This chapter concludes the results and explain the best
method to perform this research to get the best results and define the future scope of study that
explains the extent to which the research area will be explored in the work.
10 | P a g e
CHAPTER 2
LITERATURE REVIEW/BACKGROUND STUDY
2005 – Early Work on Emotion Recognition from Facial Expressions: Initial studies in emotion
recognition focused on facial expression analysis using handcrafted features like Facial Action
Coding System (FACS). These systems relied on classical machine learning methods such as
SVMs and k-NN. However, these approaches struggled with accuracy and required extensive
manual preprocessing.
2012 – Breakthrough with Deep Learning in Computer Vision: The success of deep learning
models like AlexNet in image classification sparked interest in applying convolutional neural
networks (CNNs) for facial emotion recognition. Researchers began experimenting with deep
models to learn features directly from image data, improving recognition performance
significantly.
11 | P a g e
2.2. Existing Solution
Existing solutions for seizure detection using neural networks incorporate a wide range of
techniques—from classical machine learning to advanced deep learning models. Below are some
key methodologies used in earlier research:
[Link] Machine Learning Techniques: Early approaches involved algorithms like
Support Vector Machines (SVM), k-Nearest Neighbors (k-NN), and Decision Trees. These
models relied on handcrafted features extracted from EEG signals (e.g., frequency bands,
amplitude). While they offered reasonable accuracy, they struggled with generalization across
patients and required expert-driven feature engineering.
[Link] Learning-Based Models: Recent years have seen a shift toward deep neural networks
like CNNs, RNNs, and LSTMs, which can learn directly from raw EEG data. These models have
significantly improved accuracy and adaptability. However, they demand large annotated datasets
and high computational power, making them less accessible in resource-constrained settings.
[Link] Models: Combining traditional and deep learning techniques, hybrid models aim to
balance accuracy with interpretability. For instance, handcrafted features can be used as input to a
neural network, or deep features can be fed into a traditional classifier. These models often reduce
false alarms while maintaining strong performance.
[Link] & Real-Time Systems: The integration of wearable EEG devices with embedded
neural networks has enabled real-time seizure monitoring. These solutions can alert caregivers
immediately during seizure events. Nonetheless, they face challenges like battery life, processing
delays, and signal noise in ambulatory settings.
[Link] & Edge-Based Solutions: To address real-time processing issues, researchers have
explored cloud-based systems for heavy computations and edge computing for on-device
analysis. These frameworks reduce latency and enable continuous monitoring, but they raise
concerns over data privacy, connectivity, and system reliability.
[Link] AI and Federated Learning: Emerging solutions focus on Explainable AI (XAI)
to provide transparency in seizure prediction, making it easier for medical professionals to trust
model outputs. Meanwhile, federated learning enables collaborative model training across
institutions without sharing sensitive patient data.
12 | P a g e
2.3. Bibliometric Analysis
Emotion Recognition 2018 M. Shamim To analyze EEG signals for LSTM networks
from EEG Using Deep Hossain, S. identifying emotional states using provided
Learning U. Amin, M. deep learning techniques. effective
Alsulaiman, temporal
and G. modeling of brain
Muhammad signal patterns.
A Real-Time Facial 2020 Zhang et al. To create a lightweight emotion SVM-based
Expression Recognition detection model using facial method worked
System Using CNN geometry. efficiently on
limited-resource
devices
2022 M. I. B. The paper reviews various Literature review,
Emotion Detection via
Ahmed, S. machine learning approaches for analysis of ML
Facial Landmarks and
Alotaibi, identifying pediatric epilepsy, techniques (SVM,
SVM
Atta-ur- highlighting their effectiveness, CNN, RNN,
Rahman, S. challenges, and potential Decision Trees)
Dash, M. improvements in early diagnosis
Nabil, and A. and classification.
O. AlTurk
2023 To classify To classify emotions from speech MobileNetV2
Speech-Based Emotion
emotions recordings using spectrograms enabled fast and
Detection Using Deep
from speech and neural networks. accurate emotion
Learning recordings detection with
using low resource
spectrograms usage.
and neural
networks.
Table 1.1.
Real-time human emotion detection has become an essential field in artificial intelligence,
enabling machines to understand and respond to human emotions effectively. Over the past few
13 | P a g e
years, the focus has shifted from traditional machine learning methods toward deep learning
techniques such as CNNs, RNNs, and hybrid models that allow for faster and more accurate
emotion recognition. These models process facial expressions, speech, and physiological signals
to detect emotions with minimal delay, making them suitable for real-time [Link]
research has introduced lightweight architectures like MobileNet and the use of transfer learning
to support deployment on edge devices, such as smartphones and wearables. Multi-modal emotion
detection systems that combine facial, audio, and textual data have proven to be more robust,
especially in complex real-world environments Despite these advancements, challenges such as
dataset limitations, varied emotional expressions across cultures, and the demand for high-quality,
real-time performance persist. Future developments are expected to focus on improving system
accuracy, reducing latency, and enhancing model transparency through explainable AI.
What is to be Done?
[Link] Accurate Emotion Recognition Models – Create deep learning-based systems capable
of detecting emotions from facial expressions, speech, or physiological signals in real time.
[Link] Real-Time Processing – Optimize models to work with minimal latency for live
applications such as video conferencing, virtual assistants, and surveillance.
[Link] Multi-Modal Integration – Combine visual, audio, and biometric data to improve
emotion detection accuracy and robustness.
[Link] Model Generalizability – Train and test models on diverse datasets to ensure they
perform well across various demographics and conditions.
[Link] Practical Deployment – Develop lightweight architectures suitable for edge devices
and integrate with real-world systems such as mobile apps or wearable tech.
How it is to be Done?
[Link] Collection – Use publicly available emotion datasets like FER2013, AffectNet,
RAVDESS, and DEAP for model training and evaluation.
14 | P a g e
[Link] – Perform facial alignment, noise reduction, speech enhancement, and
normalization of input signals.
[Link] Design – Use CNNs, LSTMs, or hybrid models for extracting spatial and temporal
features from real-time data streams.
[Link]-Time Integration – Implement efficient frameworks using TensorFlow Lite, ONNX, or
OpenCV to deploy models on real-time systems.
[Link] Evaluation – Evaluate using accuracy, precision, recall, F1-score, and processing
time to assess both correctness and speed.
2.6. Goals/Objectives
Goal:To design and implement a real-time human emotion detection system using deep learning
techniques, capable of accurately recognizing emotional states from facial expressions and/or
speech inputs to enhance human-computer interaction across various real-world applications.
Objectives:
[Link] collect and analyze emotion datasets: Use reliable and publicly available datasets (like
FER2013, AffectNet, RAVDESS) to train the emotion detection model.
[Link] develop a real-time emotion detection model: Build a system using deep learning
techniques (e.g., CNN, LSTM) that can recognize emotions such as happiness, anger, sadness,
surprise, fear, etc.
[Link] ensure real-time processing and low latency: Optimize model size and performance to
make it deployable on real-time platforms (e.g., mobile devices, webcams, or live video feeds).
15 | P a g e
[Link] implement a user-friendly interface: Design a basic interface or application where users
can interact with the model in real time (optional based on scope).
[Link] evaluate model performance: Assess the system using performance metrics like accuracy,
precision, recall, and processing speed under real-time conditions.
[Link] explore multi-modal inputs (optional): If possible, combine facial expression data with
voice or physiological data to improve the accuracy and reliability of emotion recognition.
16 | P a g e
CHAPTER 3
DESIGN FLOW/PROCESS
a)Input Data Modality:The system should support input from various data modalities such as
real-time video (for facial expressions) and audio (for voice-based emotion cues). This allows for
flexibility in different application scenarios and user environments.
c)Feature Extraction:Extract facial landmarks, Action Units (AUs), or embeddings using deep
CNNs like VGGFace or MobileNet. For speech signals, extract features such as pitch, tone, and
MFCCs. These features are crucial for capturing nuanced emotional cues.
e)Training and Evaluation Methodology:Train the models using labeled datasets like FER2013,
RAVDESS, or AffectNet. Apply data augmentation, k-fold cross-validation, and regularization to
improve generalization and avoid overfitting.
g) Performance Metrics:Evaluate the system using metrics like classification accuracy, precision,
recall, F1-score, and inference time. Use confusion matrices to analyze model strengths and
weaknesses across emotion classes.
17 | P a g e
3.2. Design Constraints
a) Real-Time Processing Speed: The emotion detection system must be capable of processing
input data and generating results in real time. This requires the model to operate with very low
latency, ideally under one second, to ensure immediate feedback in applications such as virtual
assistants, surveillance, or human-computer interaction.
b) Hardware Limitations: Given the deployment scenarios may include mobile devices,
embedded systems, or edge devices, the system should be optimized to function efficiently on
hardware with limited computational resources. It should not rely on high-end GPUs or excessive
memory usage for emotion detection.
e) Data Privacy and Security: As the system processes sensitive facial or audio data, strong data
privacy and security measures must be enforced. The system should comply with relevant privacy
laws and frameworks, ensuring that user information is not misused or exposed during
transmission or storage.
f) Language and Cultural Variability: Emotional expression differs across cultures and
languages. The system should be designed to generalize well across diverse demographics to avoid
bias in emotion recognition, ensuring fairness and inclusiveness.
b)Deep Learning Model Selection: Choose an architecture that balances accuracy and efficiency,
such as CNN-LSTM hybrids for feature extraction and temporal modeling.
c)Feature Selection: Prioritize features that significantly contribute to seizure detection while
reducing model complexity.
d)Scalability & Deployment: Ensure the model can run on both high-performance servers and
edge devices to enhance usability in different environments.
18 | P a g e
3.4. Design Flow
1. Data Collection: Acquire facial recordings from publicly available datasets or
hospital collaborations.
2. Preprocessing: Remove noise, normalize signals, and extract relevant features.
3. Model Training: Train deep learning models with labeled seizure and non-seizure
data.
4. Evaluation: Test model performance using predefined metrics.
5. Optimization: Fine-tune hyperparameters and optimize computational efficiency.
6. Deployment: Implement real-time inference mechanisms and integrate with user
interfaces.
Objective: Ensure face expressions signals are properly filtered and artifact-free.
Actions: Apply band-pass filtering and noise removal techniques.
Expected Outcome: Clean facial signals with minimized noise.
19 | P a g e
3.6. Implementation Plan/Methodology
a) Data Collection & Preprocessing:
Utilize datasets such as CHB-MIT and TUH EEG for model training.
Implement signal processing techniques to enhance data quality.
d) Real-time Deployment:
Implement an efficient inference engine.
Optimize the model for deployment on edge devices.
Fig 1.2 Implementation step for Real time human emotion detection.
20 | P a g e
CHAPTER 4
RESULT ANALYSIS AND VALIDATION
a) Data Collection:The first step involves gathering a robust dataset that includes facial
expressions representing various human emotions. Publicly available datasets such as FER-2013,
CK+, and JAFFE were considered for training purposes. These datasets contain thousands of
labeled facial images corresponding to primary emotional states like happiness, sadness, anger,
fear, surprise, disgust, and neutrality.
b) Preprocessing: Images from the dataset are preprocessed to ensure uniformity and quality. This
involves resizing all images to a standard resolution, converting them to grayscale for reduced
complexity, and applying techniques like histogram equalization to enhance image contrast. Facial
landmarks are also detected and aligned to standardize the orientation.
c) Feature Extraction: Facial features critical for emotion detection, such as eyes, eyebrows, lips,
and nose movement, are extracted using tools like Dlib or OpenCV’s face detection module.
Convolutional layers of neural networks are then employed to learn high-level spatial features
from the input images, eliminating the need for manual extraction.
e) Real-Time Detection System: Once the model is trained and validated, it is integrated into a
real-time emotion detection pipeline. A webcam captures live video streams which are segmented
into frames. Each frame undergoes preprocessing and is fed to the trained model, which outputs
the predicted emotion. Detected emotions are displayed on the screen in real-time alongside
bounding boxes on the user’s face.
f) Deployment: The final solution is deployed as a desktop application using Python with libraries
such as OpenCV, TensorFlow/Keras, and Tkinter for GUI support. It is designed to operate in real-
time with minimal latency. The system is tested across various lighting conditions and user
expressions to ensure robustness and adaptability.
21 | P a g e
g) Evaluation:Model performance is evaluated using accuracy, precision, recall, and F1-score.
Confusion matrices are generated to identify misclassifications. Real-time performance is also
assessed in terms of frame-per-second (FPS) rate and user feedback to measure usability.
4.2. Outcomes
22 | P a g e
23 | P a g e
CHAPTER 5
CONCLUSION AND FUTURE WORK
5.1. Conclusion
In conclusion, the development of a real-time human emotion detection system using deep learning
has been a significant and rewarding endeavor. This project demonstrates how advanced
technologies can be leveraged to interpret human emotions accurately and in real-time, offering
practical applications in healthcare, customer service, education, and human-computer interaction.
The following are the key insights from the implementation:
[Link] of System Accuracy – The deep learning models, especially Convolutional Neural
Networks (CNNs), have been successfully trained and validated on benchmark emotion datasets.
Evaluation using metrics like accuracy, precision, recall, and F1-score indicates a high level of
classification performance, confirming the model’s reliability for real-time emotion recognition.
[Link] of Deep Learning Approaches – The use of CNN-based architectures proved highly
effective in automatically learning facial features and recognizing emotions with minimal manual
intervention. Real-time performance and model responsiveness also validate the system’s
feasibility for practical deployment.
[Link]-Friendly Real-Time Deployment – The final model was successfully integrated into a
real-time application using webcam input, allowing for continuous emotion monitoring. This real-
time feedback loop is essential for interactive and adaptive systems, making the solution highly
valuable for end users.
[Link] for Future Enhancements – While the current system performs well, future
improvements may include expanding to multi-modal emotion recognition by integrating voice
and physiological data, improving detection under occlusions, and deploying the model on mobile
or embedded platforms. Collaboration with psychologists and user studies can further validate and
refine the model's real-world applicability.
24 | P a g e
5.2. Future Scope
As the development of real-time human emotion detection systems continues to evolve, there is
immense potential for expanding the utility, efficiency, and accuracy of these solutions across
various domains. The following areas highlight key directions for future enhancement and
research:
a) Enhancement of Real-Time Processing
The goal is to further minimize latency in detecting emotions to ensure instant feedback
for real-world applications.
Leveraging edge computing, optimized model compression, and hardware acceleration
(such as GPU/TPU deployment) will boost real-time responsiveness.
Integration with wearable and embedded devices such as AR/VR headsets or smart glasses
can enable continuous emotion tracking in dynamic environments.
b) Improving Model Generalizability
Future models will be trained on more diverse and inclusive datasets, covering various
ethnicities, age groups, and lighting conditions to avoid bias and ensure consistent
performance.
Transfer learning and domain adaptation techniques will be employed to ensure the model
performs reliably across unseen environments and user scenarios.
c) Explainability and Interpretability of AI Models
Advanced explainable AI techniques such as Grad-CAM, attention visualization, and
feature attribution will be integrated to make emotion recognition decisions interpretable
for developers and users alike.
Transparent models will increase trust, especially in sensitive fields such as mental health
assessment and behavioral analysis.
d) Integration with Multimodal Systems
Future iterations will explore combining facial expression data with speech, body language,
and physiological signals (e.g., heart rate or skin conductance) to improve emotion
detection accuracy.
Multimodal fusion techniques will enable more holistic emotional understanding, crucial
for applications in healthcare, education, and customer service.
e) Deployment in Real-Life Scenarios
Developing mobile apps, browser-based tools, and IoT-compatible modules will extend the
reach of emotion detection systems to remote and mobile settings.
25 | P a g e
Integrating the system with platforms such as e-learning software, telehealth tools, and
virtual assistants will enable real-world testing and usage.
f) Personalized Emotion Models
Future models will incorporate personalization techniques that adapt to individual
emotional expression patterns, leading to more accurate predictions over time.
Continuous learning algorithms and feedback loops from users will allow the model to self-
improve and tailor itself to unique emotional profiles.
g) Ethical Considerations and Privacy
Further research will explore ethical deployment, focusing on responsible data use,
consent, and emotion data anonymization.
Implementing on-device processing and privacy-preserving AI (e.g., federated learning)
will ensure data remains secure, especially in sensitive environments like mental health
tracking.
26 | P a g e
REFERENCES
[1] [ M. Happy, A. Routray, and R. Gupta, “Real-Time Facial Expression Recognition Using a
Novel Expression Recognition Technique,” IEEE Transactions on Affective Computing, vol.
6, no. 3, pp. 291–302, Jul.-Sep. 2015.
[2] A. Mollahosseini, D. Chan, and M. H. Mahoor, “Going Deeper in Facial Expression
Recognition Using Deep Neural Networks,” in Proc. IEEE Winter Conference on Applications
of Computer Vision (WACV), Lake Placid, NY, USA, 2016, pp. 1–10.
[3] Y. Li, J. Zeng, S. Shan, and X. Chen, “Occlusion Aware Facial Expression Recognition Using
CNN With Attention Mechanism,” IEEE Transactions on Image Processing, vol. 28, no. 5,
pp. 2439–2450, 2019.
[4] K. Zhang, Z. Zhang, Z. Li, and Y. Qiao, “Joint Face Detection and Alignment Using Multi-
Task Cascaded Convolutional Networks,” IEEE Signal Processing Letters, vol. 23, no. 10, pp.
1499–1503, Oct. 2016.
[5] P. Ekman and W. V. Friesen, Facial Action Coding System: A Technique for the Measurement
of Facial Movement, Palo Alto: Consulting Psychologists Press, 1978.
[6] D. G. Lowe, “Distinctive Image Features from Scale-Invariant Keypoints,” International
Journal of Computer Vision, vol. 60, no. 2, pp. 91–110, 2004.
[7] M. Abadi et al., “TensorFlow: Large-Scale Machine Learning on Heterogeneous Systems,”
2015. [Online]. Available: [Link]
[8] F. Chollet, “Keras: The Python Deep Learning Library,” 2015. [Online]. Available:
[Link]
[9] S. Koelstra, C. Muhl, M. Soleymani, J. S. Lee, and T. Pun, “DEAP: A Database for Emotion
Analysis Using Physiological Signals,” IEEE Transactions on Affective Computing, vol. 3, no.
1, pp. 18–31, Jan.-Mar. 2012.
[10] P. Viola and M. J. Jones, “Robust Real-Time Face Detection,” International Journal of
Computer Vision, vol. 57, no. 2, pp. 137–154, 2004.
27 | P a g e