Multiclass Document Classifier using BERT
DOI:
https://doi.org/10.32628/IJSRSET241127Keywords:
Classification, BERT, Contextual UnderstandingAbstract
With the rapid expansion of the internet, there has been an exponential surge in data volume, encompassing a myriad of documents laden with diverse types of information. This vast expanse includes structured and unstructured data, ranging from big data sets to formatted text and unformatted content. However, this abundance of unstructured data poses significant challenges in terms of effective management. Manual classification of this burgeoning data landscape is impractical, necessitating automated solutions. In this paper, we propose leveraging advanced machine learning techniques, particularly the BERT model, to classify documents based on contextual understanding, offering a more efficient and accurate approach to handling the data deluge.
Downloads
References
Ilkay Yelmen, Ali Gunes, and Metin Zontul on “Multi-Class Document Classification Using Lexical Ontology-Based Deep Learning” Appl. Sci. 2023, 13(10), 6139; https://doi.org/10.3390/app13106139 DOI: https://doi.org/10.3390/app13106139
Kadhim, A.I. Survey on supervised machine learning techniques for automatic text classification. Artif. Intell. Rev. 2019, 52, 273–292. [Google Scholar] [CrossRef] DOI: https://doi.org/10.1007/s10462-018-09677-1
L. Deng and J. C. Platt, "Ensemble deep learning for speech recognition," in Fifteenth annual conference of the international speech communication association, 2014 DOI: https://doi.org/10.21437/Interspeech.2014-433
W. Yin, K. Kann, M. Yu, and H. Schütze, "Comparative study of CNN and RNN for natural language processing," arXiv preprint arXiv:1702.01923, 2017.
Lai, L. Xu, K. Liu, and J. Zhao, "Recurrent convolutional neural networks for text classification," in Twenty-ninth AAAI conference on artificial intelligence, 2015. DOI: https://doi.org/10.1609/aaai.v29i1.9513
B. Ay Karakuş, M. Talo, İ. R. Hallaç, and G. Aydin, "Evaluating deep learning models for sentiment classification," Concurrency and Computation: Practice and Experience, vol. 30, no. 21, p. e4783, 2018. DOI: https://doi.org/10.1002/cpe.4783
B. Yildiz and M. Tezgider, "Improving word embedding quality with innovative automated approaches to hyperparameters," Concurrency and Computation: Practice and Experience, p. e6091, 2021. doi: https://doi.org/10.1002/cpe.6091. DOI: https://doi.org/10.1002/cpe.6091
J. Gong et al., "Hierarchical graph transformer-based deep learning model for large-scale multi-label text classification," IEEE Access, vol. 8, pp. 30885-30896, 2020. DOI: https://doi.org/10.1109/ACCESS.2020.2972751
J. Devlin, M.-W. Chang, K. Lee, and K. Toutanova, "BERT: pre- training of deep bidirectional transformers for language understanding " presented at the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Minneapolis, MN, USA., 2019.
T. B. Brown et al., "Language models are few-shot learners," arXiv preprint arXiv:2005.14165, 2020.
A. Vaswani et al., "Attention is all you need," in Advances in neural information processing systems, 2017, pp. 5998-6008.
B. Yildiz, "Reinforcement learning using fully connected, attention, and transformer models in knapsack problem solving," Concurrency and Computation: Practice and Experience, e6509, 2021. doi: https://doi.org/10.1002/cpe.6509 DOI: https://doi.org/10.1002/cpe.6509