transformer architecture

Mathematical modeling of multi-label classification of job descriptions using transformer-based neural networks

This article presents the mathematical modeling of the multi-label classification task of job description texts aimed at the automatic detection of working conditions and social benefits, which can enhance communication efficiency between employers and job seekers.  The proposed approach is based on the use of the transformer-based BERT neural network, pre-trained on a multilingual corpus.  The dataset was constructed by collecting job postings from the three largest Ukrainian job search platforms: Work.ua, Robota.ua, and Jooble.org.  The collected texts were augmented

Capabilities and Limitations of Large Language Models

The work is dedicated to the study of large language models (LLMs) and approaches to improving their efficiency in a new service. The rapid development of LLMs based on transformer architecture has opened up new possibilities in natural language processing and the automation of various tasks. However, fully utilizing the potential of these models requires a thorough approach and consideration of numerous factors.