Proceedings of the Institute for System Programming of the RAS
RUS  ENG    JOURNALS   PEOPLE   ORGANISATIONS   CONFERENCES   SEMINARS   VIDEO LIBRARY   PACKAGE AMSBIB  
General information
Latest issue
Archive

Search papers
Search references

RSS
Latest issue
Current issues
Archive issues
What is RSS



Proceedings of ISP RAS:
Year:
Volume:
Issue:
Page:
Find






Personal entry:
Login:
Password:
Save password
Enter
Forgotten password?
Register


Proceedings of the Institute for System Programming of the RAS, 2022, Volume 34, Issue 6, Pages 179–184
DOI: https://doi.org/10.15514/ISPRAS-2022-34(6)-14
(Mi tisp748)
 

Evaluation of neural models’ linguistic competence: evidence from Russian predicate agreement

K. A. Studenikina

Lomonosov Moscow State University
Abstract: This study investigates the linguistic competence of modern language models. Artificial neural networks demonstrate high quality in many natural language processing tasks. However, their implicit grammar knowledge remains unstudied. The ability to judge a sentence as grammatical or ungrammatical is regarded as key property of human’s linguistic competence. We suppose that language models’ grammar knowledge also occurs in their ability to judge the grammaticality of a sentence. In order to test neural networks’ linguistic competence, we probe their acquisition of number predicate agreement in Russian. A dataset consisted of artificially generated grammatical and ungrammatical sentences was created to train the language models. Automatic sentence generation allows us to test the acquisition of particular language phenomenon, to detach from vocabulary and pragmatic differences. We use transfer learning of pre-trained neural networks. The results show that all the considered models demonstrate high accuracy and Matthew's correlation coefficient values which can be attributed to successful acquisition of predicate agreement rules. The classification quality is reduced for sentences with inanimate nouns which show nominative-accusative case syncretism. The complexity of the syntactic structure turns out to be significant for Russian models and a model for Slavic languages, but it does not affect the errors distribution of multilingual models.
Keywords: linguistic competence, language models, transfer learning, artificial intelligence, natural language processing, grammaticality judgments
Funding agency Grant number
Intellect
This work was supported by Non-commercial Foundation for support of Science and Education «INTELLECT».
Document Type: Article
Language: Russian
Citation: K. A. Studenikina, “Evaluation of neural models’ linguistic competence: evidence from Russian predicate agreement”, Proceedings of ISP RAS, 34:6 (2022), 179–184
Citation in format AMSBIB
\Bibitem{Stu22}
\by K.~A.~Studenikina
\paper Evaluation of neural models’ linguistic competence: evidence from Russian predicate agreement
\jour Proceedings of ISP RAS
\yr 2022
\vol 34
\issue 6
\pages 179--184
\mathnet{http://mi.mathnet.ru/tisp748}
\crossref{https://doi.org/10.15514/ISPRAS-2022-34(6)-14}
Linking options:
  • https://www.mathnet.ru/eng/tisp748
  • https://www.mathnet.ru/eng/tisp/v34/i6/p179
  • Citing articles in Google Scholar: Russian citations, English citations
    Related articles in Google Scholar: Russian articles, English articles
    Proceedings of the Institute for System Programming of the RAS
     
      Contact us:
     Terms of Use  Registration to the website  Logotypes © Steklov Mathematical Institute RAS, 2025