Text classification is a natural language processing (NLP) problem that aims to classify previously unseen texts. In this study, Bidirectional Encoder Representations for Transformers (BERT) architecture is preferred for text classification. The classification is aimed explicitly at a chatbot that can give automated responses to website visitors' queries. BERT is trained to reduce the need for RAM and storage by replacing multiple separate models for different chatbots on a server with a single model. Moreover, since a pre-trained multilingual BERT model is preferred, the system reduces the need for system resources. It handles multiple chatbots with multiple languages simultaneously. The model mainly determines a class for a given input text. The classes correspond to specific answers from a database, and the bot selects an answer and replies back. For multiple chatbots, a special masking operation is performed to select a response from within the corresponding bank answers of a chatbot. We tested the proposed model for 13 simultaneous classification problems on a data set of three different languages, Turkish, English, and German, with 333 classes. We reported the accuracies for individually trained models and the proposed model together with the savings in the system resources.
BERT classification chatbot memory gain multi-domain multi-lingual.
Birincil Dil | İngilizce |
---|---|
Konular | Yapay Zeka |
Bölüm | Research Articles |
Yazarlar | |
Yayımlanma Tarihi | 30 Aralık 2021 |
Gönderilme Tarihi | 30 Kasım 2021 |
Yayımlandığı Sayı | Yıl 2021 Cilt: 1 Sayı: 2 |
All articles published by JAIDA are licensed under a Creative Commons Attribution 4.0 International License.