33rd IEEE Conference on Signal Processing and Communications Applications, SIU 2025, İstanbul, Turkey, 25 - 28 June 2025, (Full Text)
In recent years, transformer-based models pre-trained on extensive corpora have played a critical role in the advancement of Natural Language Processing methodologies. Particularly, methods based on BERT have demonstrated remarkable performance across various tasks by offering robust capabilities in deeply understanding texts semantically. However, despite these advancements, there is a notable scarcity of studies applying these technologies in the aviation sector. This paper develops a multi-class classification model for aviation-specific texts using variants of BERT. The study encompasses the processes of collecting web content related to aircraft, labeling and model training. The details of the dataset are explained and the outcomes of the study are assessed based on the macro F1-score and accuracy of different models.