Weston and Collobert (2008)
|
Convolution neural network architectures were proposed for performing NLP tasks |
Text classification process can be performed using CNN. |
Word embedding techniques were not used. |
Kim (2014)7 Kim Y. Convolutional neural networks for sentence classification. In Proceedings of the Conference on Empirical Methods in Natual Language Processing (EMNLP); 2014 Oct; Doha, Qatar. Association for Computational Linguistics; c2014. p.1746-1751.
|
Suggested to implement feature extraction techniques along with deep neural network methods. |
Simple CNN and word2vec with slight modification in the hyper parameters, yields optimal classification of NLP tasks. |
Single layer CNN is used. This will not perform efficiently for huge datasets. |
Le Quoc and T. Mikolov (2014)
|
Proposed an unsupervised algorithm named paragraph vector to overcome the drawbacks of bag-of-words. |
Paragraph vector works efficiently than bag of words in text classification |
Implementation of paragraph vector is expensive |
Pennington et al. (2014)
|
Proposed a global regression model that combines two models namely global matrix factorization and local context window method. |
The vector space produced by the model is with meaningful sub structures. |
The models performance varies based on the number of negative samples. |
Bozyigit et al. (2015)
|
Five classifiers and two feature selection methods in the Text classifications were evaluated on news dataset. |
Best classification accuracy is obtained using this combination on the dataset. |
The classification accuracy is not achieved for large datasets. |
Ming and Xianchun (2016)
|
Proposed the doc2vec model which is the combination of word2vec and clustering algorithm to express the information of document. |
TF-IDF algorithm is used along with word2vec to form document vectors. |
Single layer CNN is used which holds good for small documents. Multiple layers is missing for handling large documents. |
Andrei and Radu (2017)
|
The new approach proposed for text classification is clustering based word embeddings using k-means. |
The proposed work provides better results than bag of words approach. |
Alternate to k-means algorithm can be implemented to yield better results. |
Hughes et al. (2017)
|
The proposed approach is to perform classification at sentence level using deep convolutional neural network. |
Multilayer deep convolutional neural network generates optimal features to represent semantics. |
This approach has scalability issue. It works only for small datasets. |
Kilimci et al (2018)8 Kilimci Zeynep H,Akyokus S. Deep Learning- and Word Embedding-Based Heterogeneous Classifier Ensembles for Text Classification. Hindawi Complexity. 2018 Oct; 2018(7): 1-10.
|
Different word embeddings and ensemble learning for classifiers is proposed for text classification. |
The use of heterogeneous ensembles with word embeddings and deep learning enhances the text classification. |
Selecting the appropriate ensemble technique to yield optimal accuracy is challenging. |
Roger et al (2019)
|
Word embedding models along with machine learning models is proposed for hierarchical text classification. |
Word2vec, Glove and fastText proved to be best classification models. |
Hierarchical text classification becomes complex while handling the real time continuous data. |
Yao et al (2019)
|
Graph convolutional neural network is proposed for text classification. |
Single text graph is built for word corpus and then Text Graph convolutional network built for the corpus yields better results. |
This approach lowers the training percentage in the dataset. |
Albalawi et al. (2021)
|
Deep learning models like BiLSTM with word embeddings are compared with traditional machine learning models for health related tweets from social media. |
The classification accuracy is more with deep learning model when compared with ML models. |
The use of advanced Deep learning techniques like auto encoders may impact better than the used approaches. |
Guilherme et al. (2021)
|
An embedding technique (Distance based vector Embedding) based on Logistic Markov Embedding is proposed. |
Scalability issue is addressed using the proposed model along with negative sampling approach. |
The work limits with machine learning approaches. Deep learning techniques were not implemented. |
Moreo et al. (2021)
|
Proposed word class embedding methods were merged with pre-trained word embeddings for solving NLP tasks. |
The proposed work enhances the deep learning training and multiclass classification. |
This approach is not suitable for binary classification. |
Pittaras et al. (2021)
|
Semantics were extracted for each word and then word2vec embedding model is applied. |
Applying semantics yields better performance on text classification. |
The classification models computation complexity is increased. |