论文标题
使用变压器的孟加拉文本分类
Bangla Text Classification using Transformers
论文作者
论文摘要
文本分类一直是NLP中最早的问题之一。随着时间的流逝,应用领域的范围扩大了,处理新领域的困难(例如,嘈杂的社交媒体内容)也有所增加。解决问题的策略从经典的机器学习转变为深度学习算法。最近的深度神经网络体系结构之一是变压器。使用这种类型的网络及其变体设计的模型最近在许多下游自然语言处理任务中表现出了成功,尤其是对于资源丰富的语言,例如英语。但是,对于孟加拉文本分类任务尚未完全探索这些模型。在这项工作中,我们为孟加拉文本分类任务进行了多种语言变压器模型,包括情绪分析,情感检测,新闻分类和作者身份归因。我们在六个基准数据集中获得了最新结果的状态,从而在不同的任务中提高了5-29%的准确性。
Text classification has been one of the earliest problems in NLP. Over time the scope of application areas has broadened and the difficulty of dealing with new areas (e.g., noisy social media content) has increased. The problem-solving strategy switched from classical machine learning to deep learning algorithms. One of the recent deep neural network architecture is the Transformer. Models designed with this type of network and its variants recently showed their success in many downstream natural language processing tasks, especially for resource-rich languages, e.g., English. However, these models have not been explored fully for Bangla text classification tasks. In this work, we fine-tune multilingual transformer models for Bangla text classification tasks in different domains, including sentiment analysis, emotion detection, news categorization, and authorship attribution. We obtain the state of the art results on six benchmark datasets, improving upon the previous results by 5-29% accuracy across different tasks.