Vietnamese Sentiment Analysis Report
Vietnamese Sentiment Analysis Report
analysis
December 2023, HUST
Deep Learning
2 Dataset 2
2.1 Data set . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2
2.2 Data preproccessing . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3
2.3 TD-IDF . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4
2.4 Word Embedding (word2vec) . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5
2.4.1 One-Hot Vectors . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5
2.4.2 The Skip-Gram Model . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5
2.4.3 The Continuous Bag of Words Model . . . . . . . . . . . . . . . . . . . . 6
2.5 Exploratory Data Analysis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6
3 Methodology 7
3.1 Machine learning methods . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7
3.1.1 XGBoost . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7
3.1.2 SVM . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8
3.2 Deep learning methods . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8
3.2.1 Convolutional neural network . . . . . . . . . . . . . . . . . . . . . . . . 8
3.2.2 RNN . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9
3.2.3 LSTM . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9
3.2.4 GRU . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10
3.2.5 BERT . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10
3.2.6 XLM-RoBERTa . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12
3.2.7 PhoBERT . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12
4 Results 13
4.1 Metrics used . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13
4.2 Model selection . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13
4.3 Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15
6 References 17
1
Vietnamese sentiment analysis
Abtract
Sentiment analysis serves as a potent approach for extracting valuable insights from reviews,
categorizing them based on positive or negative sentiments. In the context of the UIT-VSMEC
(Vietnamese Sentiment Emotion) Project, our goal is to employ Machine Learning and Deep
Learning methodologies to assess the model’s accuracy and determine the optimal algorithm
for sentiment analysis.
1 Introduction
In the era of the internet and rapid technological advancement, the development of intelligent
systems has gained substantial momentum, aiming to automate tasks traditionally performed
by humans. Within the domain of Natural Language Processing (NLP), a particularly intriguing
challenge has emerged: categorizing text-based comments into distinct emotional categories.
Unlike traditional sentiment analysis, which classifies text as positive, negative, or neutral,
this project delves into the more nuanced realm of emotion recognition. Emotion analysis
plays a pivotal role in understanding the sentiment and underlying emotions expressed in text,
offering valuable insights across various applications, from assessing user reactions to content
to understanding customer feedback in depth. To tackle the task of categorizing comments
into emotions, we employ a diverse array of machine learning and deep learning models. These
models are designed to harness the power of multi-layered neural networks, allowing them to
learn intricate feature representations from the training data and develop predictive models
for emotional categorization. However, our primary focus lies on the latest deep learning
techniques, which have demonstrated remarkable capabilities in capturing the nuanced nuances
of human emotions expressed in text. Central to our project is the UIT-VSMEC (University
of Information Technology - Vietnamese Sentiment Emotion Corpus) dataset, meticulously
designed for the Vietnamese language. This dataset serves as the cornerstone for training and
evaluating our machine learning and deep learning models, providing them with a rich source
of labeled data to learn from. Our endeavor represents a pioneering effort in the realm of
emotion analysis for Vietnamese text, aimed at not only understanding the sentiment but also
decoding the intricate tapestry of emotions expressed by users. Through the fusion of cutting-
edge technology and language-specific dataset, we embark on a journey to advance emotion
recognition and bring deeper insights to the world of text-based communication.
2 Dataset
2.1 Data set
Vietnamese Social Media Emotion Corpus (UIT-VSMEC) has about 6,927 human-annotated
sentences with six emotion labels (sadness, enjoyment, anger, disgust, fear and surprise),
contributing to emotion recognition research in Vietnamese which is a low-resource language
in Natural Language Processing (NLP).
Based on Ekman’s instruction in basic human emotions, they classify for Vietnamese text
with seven emotion labels described as follows:
- Enjoyment: For comments with the states that are triggered by feeling connection or
Page 2
Vietnamese sentiment analysis
sensory pleasure. It contains both peace and ecstasy. The intensity of these states varies from
the enjoyment of helping others, a warm uplifting feeling that people experience when they see
kindness and compassion, an experience of ease and contentment or even the enjoyment of the
misfortunes of another person to the joyful pride in the accomplishments or the experience of
something that is very beautiful and amazing. For example, the emotion of the sentence "Nháy
mt thôi cũng đáng yêu, kkk" (English translation: "Just the act of winking is so lovely!") is
Enjoyment.
- Sadness: For comments that contain both disappointment and despair. The intensity
of its states varies from discouragement, distraughtness, helplessness, hopelessness to strong
suffering, a feeling of distress and sadness often caused by a loss or sorrow and anguish. The
Vietnamese sentence "Lúc đy kh lm... k nim :( " (English translation: "It was hard that
time..memory :( " ) has an emotion of Sadness, for instance.
- Fear: For comments that show anxiety and terror. The intensity of these states varies
from trepidation - anticipation of the possibility of danger, nervousness, dread to desperation,
a response to the inability to reduce danger, panic and horror - a mixture of fear, disgust and
shock. A given sentence "Chuyn này làm tao ni ht da gà" (English translation: "This story
causes me goosebumps") is a Fear labeled-sentence
- Anger: For comments with states that are triggered by a feeling of being blocked in our
progress. It contains both annoyance and fury and varies from frustration which is a response
to repeated failures to overcome an obstacle, exasperation - anger caused by strong nuisance,
argumentativeness to bitterness - anger after unfair treatment and vengefulness. For example,
"Bin m mày đi!" (English translation: "You fucking get lost!") is labeled with Angry.
- Disgust: For comments which show both dislike and loathing. Their intensity varies from
an impulse to avoid something disgusting or aversion, the reaction to a bad taste, smell, thing
or idea, repugnance to revulsion which is a mixture of disgust and loathing or abhorrence -
a mixture of intense disgust and hatred. As "Làm bn vi my th loi này nhc c ngi" (English
translation: "Making friends with such types humiliates you") has an emotion of Disgust.
- Surprise: For comments that express the feeling caused by unexpected events, something
hard to believe and may shock you. This is the shortest emotion of all emotions, only takes a
few seconds. And it passes when we understand what is happening, and it may become fear,
anger, relief or nothing ... depends on the event that makes us surprise. "Trên đi còn tn ti th
này sao??" (English translation: "How the hell in this world this still exists??") is annotated
with Surprise
- Other: For comments that show none of those emotions above or comments that do not
contain any emotions. For instance, "Mình đã xem rt nhiu video nh này ri nên thy cũng bình
thng" (English translation: "I have seen a lot of videos like this so it’s kinda normal") is neutral,
so its label is Other.
Page 3
Vietnamese sentiment analysis
comma, dots, question mark, exclamation point and remove others. We believe these punctu-
ations can provides valuable sentiment information for our models (eg: exclamation point can
express surprise)
- Remove duplicate characters in words, and words in sentences such as đeppppppp quáaa.
- replace special words in Vietnamese to their original meaning: ’ô kêi’: ’ ok ’, ’o kê’: ’ ok
’, ’kh ’:’ không ’, ’kô ’:’ không ’, ’hok ’:’ không ’,...
- Handle English words and too long words: In Vietnamese, the longest word is “nghiêng”
(inclined), which contains 7 characters, so any words that is more than 7 characters will indicate
that it is misspelled. But in our dataset, there are also many English words, which can be more
than 7 characters. To solve this problem, we use an English dictionary to filter out the English
words and remove all other words with more than 7 characters
- Handle misspelled words: In Vietnam, different regions will have different ways to pronun-
ciate. Because of that, in some regions, people often have same mistakes in pronunciating and
writing Vietnamese. For example: Many people might write “cm n” or “giá r” as “cãm n” or “giá
r”. To overcome this problem, we first analyse and choose the top words that are misspelled
and manually created a small dictionary to correct the spelling mistakes.
- Handle emoticons and emojis: Since body language and verbal tone do not translate in our
text messages or e-mails, people have developed alternate ways to convey nuanced meaning.
The most prominent change to our online style has been emoticons and emoji. Emoticons are
punctuation marks, letters, and numbers used to create pictorial icons that generally display
an emotion or sentiment. In other side, emoji are pictographs of faces, objects, and symbols.
In our datasets, there are many emoji and emoticons, and how to handle them properly is
crucial, as it may contain sentiment information we need. We have tested two cases: remove
all emoji/emoticons and replace emoji/ emoticons by words and obtain that by replacing them
by word using Emoji library, we can have better results.
2.3 TD-IDF
Term Frequency-Inverse Document Frequency (TF-IDF) is a numerical statistic that reflects
the importance of a word in a document relative to a collection of documents, often used in
natural language processing and information retrieval. It is a combination of two components:
Term Frequency (TF) and Inverse Document Frequency (IDF).
Page 4
Vietnamese sentiment analysis
The TF-IDF score for a term in a specific document is obtained by multiplying its TF and
IDF values:
T F − IDF (t, d, D) = T F (t, d) ∗ IDF (t, D)
In summary, TF-IDF is a powerful tool for evaluating the importance of terms in a document
within the context of a larger collection. It allows for the identification of key terms that are
both frequent within a document and rare across the entire document set, thus aiding in tasks
such as text mining, information retrieval, and document categorization.
• θ: all variables.
• n: size of training.
• T : number of words.
Page 5
Vietnamese sentiment analysis
• θ: all variables.
• T : number of words.
The CBOW model has a number of advantages over the skip-gram model. First, it is able to
learn more complex relationships between words, as it is able to take into account the context
of the center word. Second, it is more efficient to train, as it only needs to calculate the gradient
for the center word, rather than for all of the surrounding words.
Page 6
Vietnamese sentiment analysis
3 Methodology
3.1 Machine learning methods
3.1.1 XGBoost
XGBoost, or eXtreme Gradient Boosting, is an advanced machine learning algorithm that
has gained prominence for its exceptional performance in supervised learning tasks. It operates
within the realm of ensemble learning, a methodology that combines the predictive strength of
multiple models to enhance overall accuracy and generalization. What sets XGBoost apart is its
sequential construction of decision trees, where each subsequent tree focuses on correcting the
errors of the preceding ones. This process, known as boosting, allows XGBoost to incrementally
refine its predictions, producing a robust and accurate model.
The algorithm’s optimization hinges on a carefully crafted objective function that comprises
two crucial components: a loss function and a regularization term. The loss function quantifies
the disparity between predicted and actual values, while the regularization term helps prevent
the model from becoming overly complex and overfitting the training data. Through an itera-
tive process, XGBoost minimizes this combined objective function, striking a balance between
precision and model simplicity. One key feature that contributes to XGBoost’s popularity is
its interpretability. The algorithm provides valuable insights into the importance of each fea-
ture in the dataset, offering a clear understanding of how these features influence the model’s
Page 7
Vietnamese sentiment analysis
3.1.2 SVM
Support Vector Machines (SVM) are a versatile class of supervised learning algorithms with
a robust mathematical foundation. At their core, SVMs are designed to find the hyperplane
that best separates classes in the feature space. The algorithm achieves this by identifying
support vectors—data points crucial for defining the decision boundary. The concept of a
margin, representing the distance between the hyperplane and the nearest data point of any
class, is central to SVM. SVMs excel in scenarios where the goal is to maximize this margin,
as it leads to a more resilient and generalizable model.
One notable feature of SVM is its ability to handle non-linear decision boundaries. This
is accomplished through the use of kernel functions, such as radial basis function (RBF) or
polynomial kernels, which implicitly map input features into a higher-dimensional space where
a linear separation is feasible. Its formula:
||x−z||2
K(x, z) = e 2σ , where σ > 0
The regularization parameter (C) in SVM controls the trade-off between achieving a smooth
decision boundary and correctly classifying training data. A smaller C value encourages a
broader margin, potentially allowing for some misclassifications, while a larger C emphasizes
accurate classification, potentially leading to a narrower margin.
SVMs find application in diverse domains, including image recognition, text categorization,
and bioinformatics. Their efficacy in high-dimensional spaces, even when the number of features
exceeds the number of samples, makes SVMs particularly suitable for tasks with complex data
structures. While SVMs are powerful, their performance can be influenced by the choice of
kernel and tuning parameters, requiring careful consideration in practice. Nevertheless, SVMs
remain a foundational tool in machine learning, celebrated for their ability to handle various
types of data and produce robust decision boundaries.
Page 8
Vietnamese sentiment analysis
As part of the preprocessing pipeline, VNCoreNLP is utilized for word segmentation, ensur-
ing accurate delineation of individual words in the text. This segmentation aids in the model’s
ability to comprehend the structure and context of the input, crucial for effective classification.
Instead of image pixels, the input to most NLP tasks are sentences or documents represented
as a matrix. Each row of the matrix corresponds to one token, that is, each row is vector that
represents a token.
3.2.2 RNN
Leveraging Recurrent Neural Networks (RNNs) for multiclass Vietnamese text classification,
this method integrates PhoWord2Vec (dim = 300) embeddings and VNCoreNLP tools for word
segmentation during preprocessing. RNNs excel in capturing sequential dependencies, while
PhoWord2Vec enhances semantic understanding. The use of VNCoreNLP ensures accurate
word segmentation, collectively enabling the model to discern nuanced patterns in Vietnamese
text for classification.
3.2.3 LSTM
Long Short-Term Memory (LSTM) is a type of Recurrent Neural Network (RNN) that is
capable of learning long-term dependencies in sequence data. This is particularly useful for
tasks involving sequential inputs like natural language processing, speech recognition, and time
series prediction. LSTM models overcome the limitations of traditional RNNs, such as vanishing
gradient problem, by using a unique gating mechanism that controls the flow of information
between cells in the network.
The LSTM layers take into account not only the word order but also their contextual
significance within the sentence. Through this process, the model discerns key patterns residing
within the sequences, identifying their correlation with specific emotions. The ultimate layer of
the model typically consists of a softmax layer, producing a probability distribution of potential
emotions. With the highest-probability emotion selected as its prediction, this model possesses
a key advantage in its adaptability to varying sentence lengths due to the inherent properties
of LSTM networks. As such, it boasts immense versatility and efficacy in accurately classifying
the conveyed emotion.
Page 9
Vietnamese sentiment analysis
3.2.4 GRU
Introduced in 2014 by Kyunghyun Cho et al., GRU has established itself as a well-known
type of recurrent neural network. It was designed as a simpler alternative to Long Short-Term
Memory (LSTM) networks, with fewer parameters, making it more computationally efficient.
GRU has gained popularity in the field of deep learning, particularly in tasks involving
sequential data like natural language processing, speech recognition, and time-series prediction.
Its simpler architecture compared to LSTM makes it faster to train, which can be advantageous
in projects where computational resources or time are limiting factors.
However, it’s important to note that while GRU has been widely adopted, LSTM is still
more prevalent in certain applications, especially those that require modeling longer sequences
and more complex dependencies. The choice between GRU and LSTM often depends on the
specific requirements of the task at hand. Despite its relative simplicity, GRU has proven to
be a powerful tool in the deep learning toolkit.
3.2.5 BERT
One of the biggest challenges in natural language processing (NLP) is the shortage of train-
ing data. Because NLP is a diversified field with many distinct tasks, most task-specific datasets
contain only a few thousand or a few hundred thousand human-labeled training examples. How-
ever, modern deep learning-based NLP models see benefits from much larger amounts of data,
improving when trained on millions, or billions, of annotated training examples. To help close
this gap in data, researchers have developed a variety of techniques for training general purpose
language representation models using the enormous amount of unannotated text on the web
(known as pre-training). The pre-trained model can then be fine-tuned on small-data NLP tasks
like question answering and sentiment analysis, resulting in substantial accuracy improvements
compared to training on these datasets from scratch. Open sourced a new technique for NLP
pre-training called Bidirectional Encoder Representations from Transformers, or BERT. With
this release, anyone in the world can train their own state-of-the-art question answering system
(or a variety of other models) in about 30 minutes on a single Cloud TPU, or in a few hours
using a single GPU. The release includes source code built on top of TensorFlow and a number
of pre-trained language representation models. In our associated paper, we demonstrate state-
of-the-art results on 11 NLP tasks, including the very competitive Stanford Question Answering
Dataset (SQuAD v1.1). Input: The input data of BERT concludes 3 components which are
Page 10
Vietnamese sentiment analysis
token embedding which separates tokens in an input; segment embedding, this helps the model
to distinguish different sentences in an input; finally positional embedding which indicate the
position of each token.
MLM: The first idea used in BERT is Masked-Language Modeling (MLM). BERT tries to
predict words in a sentence which are randomly masked before by it. The masking proportion
is about 15 percent and the masked token will be replaced by token [MASKED]. BERT use the
bidirectional approach, so that it can look both previous and next tokens, understand the full
context of the sentence to predict the masked words.
NSP: The second idea in BERT is Next Sentence Prediction (NSP). This technique is used
to learn the relationship between two sentences. BERT receives the input of two sentences and
tries to predict whether the second sentence is the next sentence of the first sentence. During
training, half of time the truth second sentence is fed with the first sentence and half of time
the second sentence is a random sentence.
Page 11
Vietnamese sentiment analysis
3.2.6 XLM-RoBERTa
XLM-RoBERTa model pre-trained on 2.5TB of filtered CommonCrawl data containing 100
languages. It was introduced in the paper Unsupervised Cross-lingual Representation Learning
at Scale by Conneau et al.
XLM-RoBERTa is a multilingual version of RoBERTa. It is pre-trained on 2.5TB of filtered
CommonCrawl data containing 100 languages.
RoBERTa is a transformers model pretrained on a large corpus in a self-supervised fashion.
This means it was pretrained on the raw texts only, with no humans labelling them in any way
(which is why it can use lots of publicly available data) with an automatic process to generate
inputs and labels from those texts.
More precisely, it was pretrained with the Masked language modeling (MLM) objective.
Taking a sentence, the model randomly masks 15 percentage of the words in the input then run
the entire masked sentence through the model and has to predict the masked words. This is
different from traditional recurrent neural networks (RNNs) that usually see the words one after
the other, or from autoregressive models like GPT which internally mask the future tokens. It
allows the model to learn a bidirectional representation of the sentence. This way, the model
learns an inner representation of 100 languages that can then be used to extract features useful
for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a
standard classifier using the features produced by the XLM-RoBERTa model as inputs.
3.2.7 PhoBERT
PhoBERT, derived from RoBERTa, incorporates notable distinctions in both dataset selec-
tion and preprocessing techniques. Dat [Link] identified two primary challenges in developing a
Vietnamese language model.
Firstly, the absence of diverse monolingual training data, with only the Vietnamese Wikipedia
corpus available for training, proved limiting. This dataset, employed in the pretraining of
multilingual models like XLM-R, is constrained by its formality and relatively small size (1GB
uncompressed). Expanding the pretraining data is crucial for enhancing pretrained language
models.
Secondly, Vietnamese text employs white spaces to separate syllables within words. Tra-
ditional BERT-based models did not account for this linguistic nuance. The solution involved
augmenting the pretraining data by incorporating a second corpus ( 19GB) sourced from news
websites and topics. To address the whitespace issue, RDRSegmenter from VnCoreNLP was
deployed for word and sentence segmentation, yielding 145 million word-segmented sentences.
Subsequently, fastBPE, distinct from RoBERTa, was used to segment these sentences into sub-
word units using a 64K subword type dictionary. Regarding optimization, PhoBERT utilized
the RoBERTa implementation in fairseq, running for 40 epochs. For PhoBERT base, optimized
by Adam with a batch size of 1024, a peak learning rate of 0.0004 was employed. PhoBERT
large used a batch size of 512, a peak learning rate of 0.0002, and comprised 13 hidden states.
The parameters of PhoBERT were frozen, and the model’s input traversed PhoBERT, with
the output from some last hidden states fed into a subsequent layer. This subsequent layer
involved a 2D convolutional neural network with m output channels, followed by 2D max
pooling with a kernel size of 3. The resulting output underwent flattening, passing through a
fully connected layer and a logsoftmax activation function.
Page 12
Vietnamese sentiment analysis
4 Results
4.1 Metrics used
Let:
• T Ni true negative) is the number of instances inside ci that are assigned correctly to
another class.
• F Pi (false possitive) is the number of instances that are assigned incorrectly to classci .
• F Ni (false negative) is the number of instances inside ci that are assigned incorrectly to
another class.
• F1 - score is the harmonic mean of precision and recall. It can provide a unified view on
the performance of a classifier and is computed as:
2 ∗ P recision ∗ Recall
F1 =
P recision + Recall
Where ti is the truth label and pi is the softmax probability for the it h class.
Page 13
Vietnamese sentiment analysis
• Weight initialization: Xavier initialization. All the weights of a layer L are picked ran-
1
domly from a normal µ = 0 and variance σ 2 = L where nL is the number of neurons in
n
layer L − 1. This will prevent the gradients of network’s activations from vanishing or
exploding.
• Optimizer: Adam. It integrates the pros of both Momentum and RMSprop. It utilizes the
squared gradients to scale the learning rate as RMSprop and it is similar to the momentum
by using the moving average of the gradient. Its advantages are more memory efficient
and less computational ower.
Page 14
Vietnamese sentiment analysis
4.3 Results
Figure 4 shows the results of our models with the best embedding technique. Here, Pre
stands for pretrained embedding:
• SVM– despite being a simple model, has pretty good result with 53.54. We can see that
TF – IDF with machine learning methods still perform quite well in the sentiment analysis
task.
• CNN architecture’s performance is slightly better than GRU in this case and just lower
than LSTM in DNN architecture, but have training time much faster than LSTM.
• phoBERT, with huge improvements in dealing with Vietnamese language, has outper-
formed all other models with a big gap, outpacing the best Deep learning model we
trained (pre + XLM-roberta+ Attention) by 3% in F1 score
Figure 5 shows the detailed classification result of phoBERT model on the benchmark
dataset. We can observe that the results are quite good with 0.66 and 0.657 F1 – score,
respectively.
Page 15
Vietnamese sentiment analysis
Fig. 11: Confusion matrix of the best classification model on the UIT-VSMEC
Page 16
Vietnamese sentiment analysis
6 References
[1] NGUYEN, Dat Quoc; NGUYEN, Anh Tuan. PhoBERT: Pre-trained language models
for Vietnamese. arXiv preprint arXiv:2003.00744, 2020.
[2] DEVLIN, Jacob, et al. Bert: Pre-training of deep bidirectional transformers for language
understanding. arXiv preprint arXiv:1810.04805, 2018.
[3] DANG, Nhan Cach; MORENOGARCÍA, María N.; DE LA PRIETA, Fernando. Senti-
ment analysis based on deep learning: A comparative study. Electronics, 2020, 9.3: 483.
[4] LIU, Yinhan, et al. Roberta: A robustly optimized bert pretraining approach. arXiv
preprint arXiv:1907.11692, 2019.
[5] CUI, Zhiyong, et al. Deep bidirectional and unidirectional LSTM recurrent neural net-
work for network-wide traffic speed prediction. arXiv preprint arXiv:1801.02143, 2018.
[6] BAHDANAU, Dzmitry; CHO, Kyunghyun; BENGIO, Yoshua. Neural machine transla-
tion by jointly learning to align and translate. arXiv preprint arXiv:1409.0473, 2014.
[7] VASWANI, Ashish, et al. Attention is all you need. Advances in neural information
processing systems, 2017, 30.
[8] NGUYEN, Khang Phuoc-Quy; VAN NGUYEN, Kiet. Exploiting vietnamese social
media characteristics for textual emotion recognition in vietnamese. In: 2020 International
Conference on Asian Language Processing (IALP). IEEE, 2020. p. 276-281.
[9] L. Zhang, S. Wang, and B. Liu, “Deep learning for sentiment analysis: a survey,” WIREs
Data Mining and Knowledge Discovery, vol. 8, no. 4, pp. 1942–4795, 2018, https:// onlineli-
brary. [Link]/doi/abs/10.1002/widm.1253.
[10] T. Singh and M. Kumari, “Role of text pre-processing in twitter sentiment analysis,” Pro-
cedia Computer Science, vol. 89, pp. 549–554, 2016, [Link]
pii/S1877050916311607.
[11] HOANG, Suong N., et al. An efficient model for sentiment analysis of electronic product
reviews in Vietnamese. In: International Conference on Future Data and Security Engineering.
Springer, Cham, 2019. p. 132-142.
[12] ZHANG, Ye; WALLACE, Byron. A sensitivity analysis of (and practitioners’ guide
to) convolutional neural networks for sentence classification. arXiv preprint arXiv:1510.03820,
2015.
[13] RAFFEL, Colin, et al. Exploring the limits of transfer learning with a unified text-to-
text transformer. J. Mach. Learn. Res., 2020, 21.140: 1-67.
[14] SUN, Chi, et al. How to fine-tune bert for text classification?. In: China national
conference on Chinese computational linguistics. Springer, Cham, 2019. p. 194- 206.
[15] Vong Anh Ho, et al. Emotion Recognition for Vietnamese Social Media Text. In
Proceeding of PACLING, 2019
Page 17