A Simple and Efficient Ensemble Classifier Combining Multiple Neural Network Models on Social Media Datasets in Vietnamese
Abstract
Text classification is a popular topic of natural language processing, which has currently attracted numerous research efforts worldwide. The significant increase of data in social media requires the vast attention of researchers to analyze such data. There are various studies in this field in many languages but limited to the Vietnamese language. Therefore, this study aims to classify Vietnamese texts on social media from three different Vietnamese benchmark datasets. Advanced deep learning models are used and optimized in this study, including CNN, LSTM, and their variants. We also implement the BERT, which has never been applied to the datasets. Our experiments find a suitable model for classification tasks on each specific dataset. To take advantage of single models, we propose an ensemble model, combining the highest-performance models. Our single models reach positive results on each dataset. Moreover, our ensemble model achieves the best performance on all three datasets. We reach 86.96% of F1-score for the HSD-VLSP dataset, 65.79% of F1-score for the UIT-VSMEC dataset, 92.79% and 89.70% for sentiments and topics on the UIT-VSFC dataset, respectively. Therefore, our models achieve better performances as compared to previous studies on these datasets.
1 Introduction
The rapid development of science and technology in the world has created a vast amount of data. In particular, the growth of social networks continuously creates a huge amount of comments and posts which are valuable sources to exploit and analyze in the digital era. Text classification is a prerequisite for such works as analyzing user opinion in the network environment, filtering and removing malicious information, and detecting criminal risk. With great potential, text classification has attracted much attention from experts in the natural language processing community worldwide. In English, we easily search for a range of text classification publications in many fields. However, relatively few researches have been done on Vietnamese text. Most published articles focus on binary classification. However, a large amount of information today requires analysis in many more aspects (multi-label or multi-class). The lack of knowledge and techniques for the Vietnamese language makes us decide to conduct this research to classify multi-class text for Vietnamese social media datasets. These datasets are provided from the VLSP share-task and publications on text classification. In particular, there are various social media textual datasets such as UIT-VSMEC for emotion recognition [Ho et al., 2019] and UIT-VSFC for students’ feedback classification [Nguyen et al., 2018b] and HSD-VLSP for hate speech detection [Vu et al., 2019]. These are the datasets with multi-label and imbalance between the labels that have been published recently. They are suitable for the requirements that we would like to study.
The emergence of deep neural networks [Liu et al., 2017] and word embeddings have made text classification more efficient. Pre-trained word embeddings accurately capture semantics to assist deep learning models improve the efficiency of classification. In this study, we implement deep learning models such as CNN [Kim, 2014], LSTM [Hochreiter and Schidhuber, 1997] and their variants to solve classification problems. Besides, we implement the BERT model [Devlin et al., 2018], which is a state-of-the-art model in many natural language processing tasks in recent years. BERT is trained through the transformer’s two-dimensional context (a neural network architecture based on the self-attention mechanism to understand languages). BERT is in contrast to previous deep learning models that looked at a text sequence from left to right or combined left-to-right and right-to-left training. To improve the word representation, we create a normalized words dictionary, which helps recognize words included in pre-trained embedding but is not represented due to misspellings.
As a result, CNN model combined with fastText’s pre-trained embedding [Grave et al., 2018], has been remarkably performance on Vietnamese social media datasets. Our study also proves the efficiency of BERT on Vietnamese students’ feedback dataset. Besides, we combine single models to increase the efficiency of the classification. As a result, our ensemble model accomplishes higher results than the single model. Compared to previous studies done on the datasets, our models achieve better results.
2 Related Work
Nowadays, many organizations realize the importance of sentiment analysis for consumer’s feedback. Through this feedback, they can evaluate the quality of their services or products and devise appropriate strategies. In order to predict the genre and rating of films through viewer ratings, Varshit battu and his collaborators [Varshit et al., 2018] conducted research on viewer comment data collected from many websites. They implemented various classification methods on their dataset to evaluate the effectiveness of methods. As a result, the CNN model achieved high results in many different languages.
The detection of emotions in texts has become an essential task in natural language processing. Su et al. [Su et al., 2018] studied the text emotional recognition problem based on semantic word vector and emotional word vector of the input text. Their proposed method used the LSTM model for emotion recognition by modeling the input text’s contextual emotion evolution. They use five-fold cross-validation to evaluate the performance of their proposed method. As a result, their model achieved recognition accuracy of 70.66% better than the CNN-based method which was implemented in the same dataset.
In addition, hate speech detection is increasingly concerned because of the explosion of social networks. There has been an amount of successful research in this field. To complete the offensive task of categorizing tweets that were announced by SemEval competition in 2019. Nikolov and Radivchev. [Nikolov and Radivchev, 2019] used different approaches and models towards offensive tweet classification. Their paper presented pre-processing data methods for tweets as well as techniques for tackling imbalanced class distribution in the provided test data. Their experiments show that the BERT model proved its outstanding advantages in text classification. Not only did it outperform conventional models on the validation set, but also based on the results from the test set, it did not cause the over-fitting issue.
In Vietnam, there have been some studies efforts for text classification tasks, as well as contributing Vietnamese data for the research community. Pham et al. [Pham et al., 2017] announced a neural network-based toolkit namely NNVLP for essential Vietnamese language processing tasks, including part-of-speech tagging, chunking, named entity recognition. This toolkit achieved state-of-the-art results on these three tasks. With the two of UIT-VSMEC [Ho et al., 2019] and UIT-VSFC [Nguyen et al., 2018a] datasets we used in this study, their authors performed classification tasks using a variety of deep learning methods. On the UIT-VSMEC dataset, Ho et al. [Ho et al., 2019] used Random Forest, SVM, LSTM, and CNN models to classify emotions of comments. They achieved 59.74% with seven labels and 66.48% with six labels by using the CNN model. With the UIT-VSFC dataset, Nguyen et al. [Nguyen et al., 2018a] gained the highest result by the BiLSTM model with 92.03% on sentiment and 89.62% on the topic label.
The above studies have shown the superiority of the deep learning models in text classification, which is the premise for us to apply them to the Vietnamese datasets in this study. By modifying the models and implementing new models, we aim to bring better results for these Vietnamese social media datasets.
3 Datasets
In this task, we conducted experiments on classification methods on three Vietnamese datasets, including UIT-VSMEC [Ho et al., 2019], UIT-VSFC [Nguyen et al., 2018b], and HSD-VLSP [Vu et al., 2019]. UIT-VSMEC [Ho et al., 2019] is provided by Ho et al, the items in this dataset are the Vietnamese’s comments from social networks. This dataset contains exactly 6,927 emotion annotated sentences with seven emotion labels: ENJOYMENT, SADNESS, ANGER, SURPRISE, FEAR, DISGUST, and OTHER. UIT-VSFC is provided by Nguyen et al. [Nguyen et al., 2018b]. This dataset was constructed from students’ responses from a university for the sentiment classification task. This dataset includes 16,175 items with two classification parts that are important: sentiments and topics. Each sample of the training dataset is assigned one of three sentiment labels POSITIVE, NEGATIVE, or NEURAL and assigned one of four topics’ labels LECTURERS, TRAINING PROGRAM, FACILITIES, OTHER. The HSD-VLSP dataset is a Vietnamese comments dataset about Hate Speech Detection on social networks provided by the VSLP 2019 shared-task [Vu et al., 2019]. This dataset contains the comments and posts on Facebook social networks, including 25,431 items. Each data line of the training dataset is assigned one of three labels CLEAN, OFFENSIVE, or HATE. Table 1 shows examples for each dataset and its characteristics.
Datasets Labels Percentage (%) Examples Vietnamese sentences English sentences UIT-VSMEC ENJOYMENT 28.36 Tốt quá! Very good! SADNESS 19.31 Nay buồn! Today, I feel sad! ANGER 16.59 Im đi! Shut up! SURPRISE 6.92 Đẹp trai mà bị đá à! You are handsome, but you broke up! FEAR 5.70 Tao sợ thật sự. I’m really scared. DISGUST 4.46 Diễn viên già vãi -.- The actor is so old -.- OTHER 18.66 Có thể bán lại cho tớ không? Could you sell it for me? UIT-VSFC Sentiments POSITIVE 49.80 Slide, giáo trình đầy đủ. Slide and syllabus are full. NEGATIVE 45.80 Thầy chép bảng nhiều. The lecturer writes on the board a lot. NEUTRAL 4.40 Tăng cường thiết bị.. Strengthen equipment. Topics LECTURER 71.70 Thầy chép bảng nhiều. The lecturer writes on the board a lot. PROGRAM 18.70 Slide, giáo trình đầy đủ. Slide and syllabus are full. FACILITY 4.20 Tăng cường thiết bị. Strengthen equipment. OTHER 4.70 Hài lòng về tất cả. Satisfied about it all. HSD-VLSP CLEAN 91.49 Hôm nay trời nắng đẹp. It is sunny today. OFFENSIVE 5.02 vkl. Cuss. HATE 3.49 Nói đến vậy mà cũng không hiểu. Đồ ngu. Saying that without understanding. Idiot.
4 Methodology
In this task, we use several pre-processing techniques and deep learning models (CNN [Kim, 2014], LSTM [Hochreiter and Schidhuber, 1997] their variants, and BERT [Devlin et al., 2018]) to apply on the datasets. Each model has its own strength on each label. Therefore, after implementing single models, we propose a simple and efficient ensemble approach combining the best neural network models to improve the classifier performance. Figure 1 shows an overview of the ensemble model for Vietnamese social media text.

4.1 Pre-processing
We implement several pre-processing techniques for Vietnamese comments/posts, as follows: All words are converted to lowercase. URLs and non-alphabetic characters are removed (includes number, excess whitespace). Also, we use a dictionary to convert abbreviations and slang words to normal to increase the performance of pre-trained embedding as well as use vnTokenizer to tokenize words. (Word segmentation is an important task that helps models achieve better results, especially for the Vietnamese language. More and more research works on this; these works use different methods and achieve high results, such as using CRFs and SVMs [Nguyen et al., 2006], Hybrid Approach [Huyen et al., 2008]. In this study, we use vnTokenizer [Huyen et al., 2008] to tokenize words. Table 2 shows some examples in our dictionary.
No. | Abbreviation | Vietnamese meaning |
---|---|---|
1 | “Chờiiii” | “Trời” |
2 | “vklllll” | “vkl” |
3 | “chetme” | “chết mẹ” |
4 | “kbh” | “không bao giờ” |
4.2 Word Embeddings
In this task, we use three pre-trained embeddings for the Vietnamese language, which are currently assessed as the best performance embeddings. Those are fastText [Grave et al., 2018], Word2vec [Vu, 2016] with 300 dim and 400 dim. fastText [Grave et al., 2018] is used for many languages and brought good results on social media data. Therefore, we use fastText [Grave et al., 2018] for the Vietnamese language in this task. Word2vec [Vu, 2016] is Word2vec model for the Vietnamese language and is trained on Vietnamese texts. Because they are Word2vec for the Vietnamese language, thus recognizing vast Vietnamese words and brings quite good results on datasets.
4.3 Convolutional Neural Network (CNN)
In this task, we use CNN [Kim, 2014] to classify emotions for Vietnamese text. CNN is an advanced deep learning model, which includes hidden layers such as pooling layers, convolutional layers, fully connected layers, and normalization layers. CNN achieved good results for this document classification task in general and the best on UIT-VSMEC [Ho et al., 2019] and HSD-VLSP [Luu et al., 2020].
4.4 Long Short-Term Memory (LSTM) and Variants
Like CNN [Kim, 2014], LSTM [Hochreiter and Schidhuber, 1997] is also a modern classification method. This method is strong in classification problems, and most of it has achieved high-performance classification results. Therefore, in this task, we decide to choose it to compare with other classification models. LSTM is a special kind of RNN [Medsker and Jain, 1999]. LSTM’s network architecture includes memory cells and ports that allow the storage or retrieval of information. We also use Bi-LSTM with Bidirectional [Schuster and Paliwa, 1997], BiLSTM can learn more contextual information extracted from two directions.
4.5 Gated Recurrent Units (GRU)
GRU is a variant of RNN and also achieves high results in classification problems. GRU has only two gates, that is the reset gate and the update gate. GRU does not have memory cells like LSTM [Hochreiter and Schidhuber, 1997]. It has only the outputs to make decisions and to inform the next steps. These two gates filter the information from the inputs of the cell and provide an output that satisfies both the criteria of storage past information and the ability to make current decisions most accurately.
4.6 BERT (Bidirectional Encoder Representations for Transformers)
BERT [Devlin et al., 2018] is a transfer learning model that achieves state-of-the-art results on natural language processing tasks. As opposed to directional models, which read the text input sequentially (left-to-right or right-to-left), the Transformer encoder reads the entire sequence of words at once. The model can learn the context of a word based on all of its surroundings. Fortunately, BERT is provided with a version for multilingual, including Vietnamese. We apply a BERT-based classifier with a pre-trained multilingual representation into the social media datasets in Vietnamese.
4.7 Our Proposed Ensemble Method
Ensemble approaches were very effective in previous studies [Huynh et al., 2020, Tran et al., 2020]. We also find that each model has a high classification performance on certain labels so we create an ensemble model based on our implemented neural network models. For each sentence in the test set, the ensemble model selected the label according to the different models’ votes. In the case of equal votes, the label was selected based on the model had the best classify performance for that label. Algorithm 1 shows the steps of the ensemble model.
Input: A social media text and m top-performance models ().
Output: Returning the label such that our proposed ensemble method predicts.
5 Experiments
5.1 Evaluation Metric
To make comparisons with previous studies, we used the corresponding measure based on previous studies’ measure, including Weighted F1-score for UIT-VSMEC [Ho et al., 2019], Micro F1-score for UIT-VSFC [Nguyen et al., 2018b], with HSD-VLSP dataset, the organization has stopped providing a test set to ensure fairness in the next competition, so when the contest ended, we could not submit the predicted results to test the accuracy. When doing this study because of the lack of test set, we decide to use the k-fold cross-validation (k = 5) method and Macro F1-score [Luu et al., 2020] to evaluate the models we applied to HSD-VLSP.
5.2 Experimental Settings
In this study, we experiment with deep learning models that are evaluated as superior in the field of text classification, including LSTM, CNN, BERT, and their variants for the datasets. First, we clean the input data by removing expressions, numbers, special characters, and all words are converted to lowercase. Second, we use some pre-trained embedding that published for the Vietnamese language to represent words into vectors before putting them into deep learning models, including Word2vec and fastText’s pre-trained embedding. In this step, with pre-trained embedding was trained on word-separated data, we separate the words for the datasets by the vnTokenizer [Huyen et al., 2008]. fastText’s pre-trained embedding demonstrates a more efficient representation of words with social datasets. However, we realize that there are some specific words to the classification classes, although represented in pre-trained embedding, they are abbreviated or written in slang, so they can not be recognized and ignored. Therefore, we try to optimize the performance of pre-trained embedding by creating word dictionaries to normalize the above words to the normal form so that they can be represented correctly. Whether it would improve our models’ performance? The example of our dictionaries is presented in Table 2.
We also experiment with different values of parameters around the average value of sentences’ length. To evaluate whether the pre-processing of data has a significant effect on the classification results, we perform experiments on the processed data and the original data except for the UIT-VSFC dataset because it is cleaned before labeling. However, for the UIT-VSMEC, the pre-processing do not improve the efficiency compared to the original data. Therefore, we only show the results of the models on the original data for UIT-VSFC and UIT VSMEC dataset and results of the models on the pre-processing data for HSD-VLSP dataset in this study. After experimenting with single models, we combine our models using the max-voting method and model priority as presented in Algorithm 1.
5.3 Experimental Results
Our experiments achieve positive results, creating dictionaries that enhance our single model’s performance 2%. With our single models, CNN model has shown its outstanding performance when combining with fastText’s pre-trained embedding, which achieves the best results on social media commentary datasets. With the UIT-VSMEC dataset, the CNN model with three layers reach the best efficiency. Conversely, with the HSD-VLSP dataset, the CNN model has the best performance with five layers due to its larger size. Besides, BERT has higher performance than the other models when applying to the UIT-VSFC dataset. With a combination of the single models’ strengths, our ensemble model has achieved high results. Table 3 shows our results through experiments performed.
Model | UIT-VSMEC | UIT-VSFC | HSD-VLSP | ||
---|---|---|---|---|---|
Seven labels | Six labels | Sentiments | Topics | ||
BiLSTM + fastText (300 dim) | 56.93 | 64.18 | 91.53 | 88.12 | 85.19 |
BiLSTM + Word2vec (300 dim) | 56.16 | 62.24 | 90.84 | 88.53 | 78.99 |
BiLSTM + Word2vec (300 dim) | 56.40 | 60.56 | 91.18 | 88.25 | 79.18 |
LSTM + fastText (300 dim) | 52.12 | 59.02 | 91.06 | 87.42 | 85.56 |
LSTM + Word2vec (300 dim) | 50.69 | 57.27 | 90.90 | 86.79 | 76.26 |
LSTM + Word2vec (400 dim) | 49.32 | 56.58 | 90.58 | 87.36 | 80.84 |
GRU + fastText (300 dim) | 52.88 | 59.45 | 90.77 | 88.02 | 84.95 |
GRU + Word2vec (300 dim) | 54.02 | 59.70 | 90.74 | 87.71 | 77.84 |
GRU + Word2vec (300 dim) | 54.30 | 59.75 | 90.87 | 87.68 | 78.38 |
CNN + fastText (300 dim) | 59.87 | 66.54 | 90.42 | 87.68 | 85.74 |
CNN + Word2vec (300 dim) | 32.65 | 33.76 | 89.16 | 85.94 | 79.01 |
CNN + Word2vec (400 dim) | 23.76 | 47.33 | 89.79 | 86.16 | 81.63 |
BERT (Based-multilingual-cased) | 49.22 | 55.97 | 92.51 | 89.67 | 65.11 |
Our proposed model | 65.79 | 70.99 | 92.79 | 89.70 | 86.96 |
5.4 Comparison With Previous Studies
The results of our single models as well as the ensemble model are better than those of previous studies that were conducted on the same data set. We follow the same metrics and test set that previous studies have used to make similar comparisons. Table 4, 5 and 6 show the best results we achieved compared to previous studies.
Model | F1-score (%) | |
---|---|---|
Seven labels | Six labels | |
CNN + Word2vec [Ho et al., 2019] | 59.74 | 66.34 |
CNN + fastText (Our implementation) | 59.87 | 66.54 |
Our proposed ensemble (GRU + CNN + BiLSTM + LSTM) | 65.79 | 70.99 |
Model | F1-score (%) | |
---|---|---|
Sentiments | Topics | |
BiLSTM + Word2vec [Nguyen et al., 2018a] | 92.03 | 89.62 |
LD + SVM [Nguyen et al., 2018c] | 92.20 | - |
BERT (Our implementation) | 92.51 | 89.67 |
Our proposed ensemble (BERT + CNN + BiLSTM + LSTM) | 92.79 | 89.38 |
Our proposed ensemble (BERT + CNN + BiLSTM) | 92.13 | 89.70 |
Model | F1-score (%) |
---|---|
Text-CNN [Luu et al., 2020] | 83.04 |
Logistic regression∗ [Pham et al., 2019] | 61.97 |
Logistic regression + Random Forest + Extra Tree∗ [Dang et al., 2019] | 58.88 |
VDCNN, TextCNN, LSTM, LSTMCNN, SARNN∗ [Nguyen et al., 2019] | 58.45 |
BiLSTM∗ [Do et al., 2019] | 56.28 |
CNN + fastText (Our implementation) | 85.74 |
Our proposed ensemble (CNN + BiLSTM + LSTM) | 86.96 |
6 Result Analysis
Through the experiment on the UIT-VSMEC dataset, we find that the classification model got a high accuracy for SADNESS, and ENJOYMENT labels. The labels ENJOYMENT and SADNESS are the two labels that account for a high proportion in the dataset, so the correct classification rate for these labels is quite high. However, each model has its strengths in labels. CNN accomplish the highest accuracy classification rate on ENJOYMENT (73.70%) and is relatively accurate for the rest. BiLSTM has the best effect at DISGUST (65.90%) and SADNESS (63.79%) label. Although LSTM and GRU do not have superior results compared to other models on each label, they play an essential role in voting in the ensemble model. The ensemble model improves the predictions in the ENJOYMENT label from 73.70% to 75.64%, the SADNESS label from 63.79% to 67.20%, and from 63.04% to 71.73% for the FEAR label.
In the UIT-VSFC dataset, with the sentiments field, we find that the classifying model achieve a high right classifying rate on the NEGATIVE and POSITIVE label for the sentiment classification task. These two labels occupy approximately the same rate and are much higher than the NEUTRAL label. Therefore, when classifying, the NEUTRAL label is mistaken for the two, the classification results on the NEUTRAL label are low. With sentences that are too long, sentences that express many emotions in the sentence or sentences that are too short, lacking in subjects makes the model difficult to recognize the feelings of the sentence. The highest accuracy of the NEGATIVE and POSITIVE label is 95.95% and 94.71% when implementing BERT model. BiLSTM model achieves the highest performance for NEUTRAL label at 41.31% while CNN and LSTM have the classification results almost equal to the remaining models. Our ensemble model improves the accuracy of the NEGATIVE label from 95.95% to 97.01%.
For the topic classification task, correct classification was high on the LECTURER and FACILITY labels. The LECTURER label has the highest rate in the dataset with 71.70%, so the exact classification model is easy to explain. The LECTURER label and FACILITY label have the highest accuracy by using the BiLSTM model with 95.72% and 93.79%. The highest performance of PROGRAM and OTHER label is 81.29% and 48.42% when implementing BERT. The FACILITY label only accounts for the lowest rate in the dataset with 4.30%. This label contains words related to tools, facilities in the school. Hence, it was easy to identify and difficult to be confused with other labels. Although FACILITY label took up a small percentage, this label reached a high rate of correct categorization. The OTHER label is the one with the lowest accuracy rating because it is difficult to distinguish from the others. Our ensemble model does not increase the classification performance of each label superior to the single model, but it increases the model’s overall prediction rate.
In the HSD-VLSP dataset, we find that the CLEAN label achieve the highest classification result, which is also the highest percentage in the dataset (accounting for 97.82%). The other two labels tend to be mistaken for the CLEAN label due to the effect of CLEAN was too high. The words that appear a lot in the CLEAN label, the classification model would default to their positive meaning, so when using these words in other meaningful sentences (hate, offensive), the model tend to categorize them into a positive sense. Moreover, HATE and OFFENSIVE labels show the relative same emotional state, which is confusing in the classification process, resulting in a low correct classification rate on these two labels. Through each fold, the classification results of the models change. In particular, the CNN model brings the most stable accuracy through each fold, so overall through five folds this model achieves the highest accuracy. CNN accomplish the highest accuracy classification rate on CLEAN at 99.42% and OFFENSIVE at 68.60%. The HATE label has the highest accuracy by using the LSTM model with 85.10%. Through each fold, LSTM shows superiority when classifying on HATE labels, so when ensemble, we give priority to the LSTM model to classify on label HATE. Therefore, the ensemble model helps to increase the accuracy on this label from 85.10 % to 85.39 %. Besides, the ensemble model makes the model’s overall prediction rate increase.
7 Conclusion and Future Work
In this study, we propose a simple but effective ensemble method for social media text on the three different datasets: HSD-VLSP, UIT-VSFC, and UIT-VSMEC. These datasets are multi-label datasets with various labels depending on each dataset. Our ensemble model accomplish outstanding results higher than the deep neural network, as well as previous studies. The combination of many different models’ strengths has achieved higher results than single models. In UIT-VSMEC, we achieve 65.79% of F1-score. With the UIT-VSFC dataset, our results are 92.79% in sentiment classification task and 89.70% in the topic classification task in terms of F1-score. Finally, in HSD-VLSP dataset, we reach 86.96% of the F1-score.
Besides, our single models also bring better results than previous studies. With single models, we find that, with small datasets, pre-processing does not bring expected results, the UIT-VSMEC dataset gain better performances when skipping the pre-processing step. On the contrary, with large datasets, pre-processing data has a good effect on our models, like data from the VLSP shared task. Also, with specific data domains, pre-trained word embeddings on the corresponding data domain will be more effective. Besides, improving word representation by dictionaries also helps to improve the efficiency of the models. Our dictionaries improve pre-trained embedding performance and increase 2% the effectiveness of our single models. This study indicate that the BERT model has been highly effective for Vietnamese language classification. However, with the use of pre-trained word embeddings training on Wikipedia data, it does not achieve the expected results when applying to the social media data like UIT-VSMEC and UIT-VSFC. With the datasets related to Vietnamese comments on social media, the CNN model has outperformed other single models. Choosing the number of layers is essential, with small datasets such as the UIT-VSMEC dataset, enhancing layers does not bring high efficiency but only makes the model more complicated. In contrast, the addition of layers reach excellent performance for large datasets such as the HSD-VLSP dataset.
For future works, it is necessary to improve the performance of the task, obtaining better results on the datasets, by studying other neural-based models. Besides, test each model with many different parameters to find more suitable parameters. It is implementing experiments of various data processing methods as well as improving coverage of word embeddings on the datasets to gain better results.
References
- [Allen and David, 1974] Allen, David M The Relationship between Variable Selection and Data Agumentation and a Method for Prediction. Technometrics, pages 125–127.
- [Varshit et al., 2018] Battu Varshit, Batchu Vishal, Gangula Rama Rohit Reddy, Dakannagari Mohana Murali Krishna Reddy and Mamidi Radhika 2018. Predicting the Genre and Rating of a Movie Based on its Synopsis. PACLIC 32. 56. 32nd Pacific Asia Conference on Language, Information and Computation.
- [Devlin et al., 2018] Devlin Jacob, Chang Ming-Wei, Lee Kenton, Toutanova Kristina 2018. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805
- [Grave et al., 2018] Grave Edouard, Bojanowski Piotr, Gupta Prakhar, Joulin Armand, Mikolov Tomas 2018. Learning Word Vectors for 157 Languages. Proceedings of the International Conference on Language Resources and Evaluation (LREC 2018).
- [Do et al., 2019] Hang Thi-Thuy Do, Huy Duc Huynh, Kiet Van Nguyen, Ngan Luu-Thuy Nguyen and Anh Gia-Tuan Nguyen 2019. Hate Speech Detection on Vietnamese Social Media Text using the Bidirectional-LSTM Model. The Sixth International Workshop on Vietnamese Language and Speech Processing VLSP 2019.
- [Hochreiter and Schidhuber, 1997] Hochreiter Sepp and Schmidhuber Jürgen 1997. Long short-term memory. MIT Press, Neural computation, 9, 8, pages 1735-1780.
- [Pham et al., 2019] Huu Quang Pham, Trung Son Nguyen and Pham Hoang Anh 2019. Automated Hate Speech Detection on Vietnamese Social Networks The Sixth International Workshop on Vietnamese Language and Speech Processing VLSP 2019.
- [Huyen et al., 2008] Huyen Nguyen Thi Minh, Roussanaly Azim, Vinh Ho Tuong 2008. A hybrid approach to word segmentation of Vietnamese texts. International Conference on Language and Automata Theory and Applications.
- [Nguyen et al., 2018b] K. V. Nguyen, Vu Duc Nguyen, P. X. V. Nguyen, T. T. T. Hong, N. L. Nguyen 2018. Uit-vsfc: Vietnamese students’ feedback corpus for sentiment analysis. 2018 10th International Conference on Knowledge and Systems Engineering (KSE), pages 19-24.
- [Kim, 2014] Kim Yoon 2014. Convolutional neural networks for sentence classification. Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP)., pages 1746-1751
- [Liu et al., 2017] Liu Jingzhou, Chang Wei-Cheng, Wu Yuexin and Yang Yiming 2017. Deep learning for extreme multi-label text classification. Proceedings of the 40th International ACM SIGIR Conference on Research and Development in Information Retrieval, pages 115-124.
- [Schuster and Paliwa, 1997] M. Schuster and K. K. Paliwa 1997. Bidirectional recurrent neural networks. IEEE Transactions on Signal Processing.
- [Medsker and Jain, 1999] Medsker Larry, Jain Lakhmi C 1999. Recurrent neural networks: design and applications. CRC press.
- [Nikolov and Radivchev, 2019] Nikolov Alex and Radivchev Victor 2019. Nikolov-Radivchev at SemEval-2019 Task 6: Offensive Tweet Classification with BERT and Ensembles. Proceedings of the 13th International Workshop on Semantic Evaluation, pages 691-695.
- [Nguyen et al., 2006] Nguyen Cam-Tu, Nguyen Trung-Kien, Phan Xuan-Hieu, Le Nguyen Minh, Ha Quang Thuy 2006. Vietnamese word segmentation with CRFs and SVMs: An investigation. Proceedings of the 20th Pacific Asia Conference on Language, Information and Computation.
- [Nguyen et al., 2018a] P. X. V. Nguyen, T. T. T. Hong, K. V. Nguyen, N. L. Nguyen 2018. Deep Learning versus Traditional Classifiers on Vietnamese Students’ Feedback Corpus. 2018 5th NAFOSTED Conference on Information and Computer Science (NICS), pages 75-80.
- [Pham et al., 2017] Pham Thai-Hoang, Pham Xuan-Khoai, Nguyen Tuan-Anh, Le-Hong Phuong 2017. Nnvlp: A neural network-based vietnamese language processing toolkit. Proceedings of The 8th International Joint Conference on Natural Language Processing.
- [Luu et al., 2020] Son T Luu, Hung P Nguyen, Kiet Van Nguyen and Ngan Luu-Thuy Nguyen 2020. Comparison Between Traditional Machine Learning Models And Neural Network Models For Vietnamese Hate Speech Detection. IEEE RIVF 2020 Confererence
- [Huynh et al., 2020] Van Huynh, Tin, Kiet Van Nguyen, Ngan Luu-Thuy Nguyen and Anh Gia-Tuan Nguyen 2020. Job prediction: From deep neural network models to applications. 2020 RIVF International Conference on Computing and Communication Technologies (RIVF), pages 1-6.
- [Su et al., 2018] Su Ming-Hsiang, Wu Chung-Hsien, Huang Kun-Yi and Hong and Qian-Bei 2018. LSTM-based text emotion recognition using semantic and emotional word vectors.. 2018 First Asian Conference on Affective Computing and Intelligent Interaction (ACII Asia).
- [Nguyen et al., 2019] Thai Binh Nguyen and Quang Minh Nguyen and Thu Hien Nguyen and Ngoc Phuong Pham and The Loc Nguyen and Quoc Truong Do 2019. VAIS Hate Speech Detection System: A Deep Learning based Approach for System Combination. Association for Vietnamese Language and Speech Processing 2019
- [Nguyen et al., 2018c] V. D. Nguyen, K. V. Nguyen and N. L. Nguyen 2018. Variants of Long Short-Term Memory for Sentiment Analysis on Vietnamese Students’ Feedback Corpus. 2018 10th International Conference on Knowledge and Systems Engineering (KSE), pages 306-311.
- [Dang et al., 2019] Van Thin Dang, Lac Si Le and Ngan Luu-Thuy Nguyen 2019. NLP@ UIT: Exploring Feature Engineer and Ensemble Model for Hate Speech Detection at VLSP 2019. The Sixth International Workshop on Vietnamese Language and Speech Processing VLSP 2019.
- [Ho et al., 2019] Vong Anh Ho, Duong Huynh-Cong Nguyen, Danh Hoang Nguyen, Linh Thi-Van Pham, Duc-Vu Nguyen, Kiet Van Nguyen, Ngan Luu-Thuy Nguyen 2019. Emotion Recognition for Vietnamese Social Media Text. PACLING 2019.
- [Vu et al., 2019] Vu Xuan-Son, Vu Thanh, Tran Mai-Vu, Le-Cong Thanh, Nguyen Huyen T M 2019. HSD shared task in VLSP campaign 2019: Hate speech detection for social good. Proceedings of VLSP 2019.
- [Vu, 2016] Xuan-Son Vu 2016. Pre-trained Word2vec models for Vietnamese. https://github.com/sonvx/word2vecVN.
- [Tran et al., 2020] Khiem Vinh Tran, Hao Phu Phan, Kiet Van Nguyen and Ngan Luu-Thuy Nguyen 2020. UIT-HSE at WNUT-2020 Task 2: Exploiting CT-BERT for Identifying COVID-19 Information on the Twitter Social Network. The 6th Workshop on Noisy User-generated Text (W-NUT) 2020.