A novel bidirectional long short-term memory model with multi-head attention for accurate language detection
dc.contributor.author | Toklu, Sinan | |
dc.contributor.author | Kabakus, Abdullah Talha | |
dc.date.accessioned | 2025-10-11T20:47:59Z | |
dc.date.available | 2025-10-11T20:47:59Z | |
dc.date.issued | 2025 | |
dc.department | Düzce Üniversitesi | en_US |
dc.description.abstract | Language detection, one of the most important elements used in natural language processing, is used extensively in various applications such as machine translation, sentiment analysis, and information retrieval. Thanks to language detection, communication between people in many different countries is possible. In addition, human-animal interaction can also be carried out in this area. In this paper, a novel Bidirectional Long Short-Term Memory model with Multi-Head Attention mechanism is proposed to accurately classify text into 17 languages, namely Arabic, Danish, Dutch, English, French, German, Greek, Hindi, Italian, Kannada, Malayalam, Portuguese, Russian, Spanish, Swedish, Tamil, and Turkish. A publicly available dataset consisting of 10,337 texts written in the above-mentioned languages is utilized to train and evaluate the proposed model. The proposed novel model achieved an extraordinary accuracy, precision, recall, and F1-score of 99.9%, outperforming the state-of-the-art baseline models. In particular, the proposed model demonstrated perfect precision (100%) for 15 languages, namely Arabic, Dutch, English, French, German, Greek, Hindi, Italian, Kannada, Malayalam, Portuguese, Russian, Swedish, Tamil, and Turkish. This research highlights the effectiveness of deep learning techniques in language detection, providing promising avenues for further advances in the field of multilingual text processing. | en_US |
dc.identifier.doi | 10.17341/gazimmfd.1543854 | |
dc.identifier.issn | 1300-1884 | |
dc.identifier.issn | 1304-4915 | |
dc.identifier.issue | 3 | en_US |
dc.identifier.scopus | 2-s2.0-105013632417 | en_US |
dc.identifier.scopusquality | Q2 | en_US |
dc.identifier.uri | https://doi.org/10.17341/gazimmfd.1543854 | |
dc.identifier.uri | https://hdl.handle.net/20.500.12684/21677 | |
dc.identifier.volume | 40 | en_US |
dc.identifier.wos | WOS:001569394800039 | en_US |
dc.identifier.wosquality | Q3 | en_US |
dc.indekslendigikaynak | Web of Science | en_US |
dc.indekslendigikaynak | Scopus | en_US |
dc.language.iso | tr | en_US |
dc.publisher | Gazi Univ, Fac Engineering Architecture | en_US |
dc.relation.ispartof | Journal of the Faculty of Engineeringand Architecture of Gazi University | en_US |
dc.relation.publicationcategory | Makale - Uluslararası Hakemli Dergi - Kurum Öğretim Elemanı | en_US |
dc.rights | info:eu-repo/semantics/openAccess | en_US |
dc.snmz | KA_WOS_20250911 | |
dc.subject | Language detection | en_US |
dc.subject | language classification | en_US |
dc.subject | translation | en_US |
dc.subject | deep learning | en_US |
dc.subject | long short-term memory | en_US |
dc.title | A novel bidirectional long short-term memory model with multi-head attention for accurate language detection | en_US |
dc.type | Article | en_US |