Supercomputing Frontiers and Innovations https://superfri.susu.ru/index.php/superfri <table cellspacing="4" cellpadding="4"> <tbody> <tr> <td style="width: 70%;" rowspan="2" align="left" valign="top"> <h3>An International Open Access Journal</h3> <p><strong>Editors-in-Chief:</strong></p> <p>Jack Dongarra, University of Tennessee, Knoxville, USA</p> <p>Vladimir Voevodin, Moscow State University, Russia</p> <p><a href="https://superfri.org/index.php/superfri/about/#custom-0"><strong>Editors-in-Chief Foreword</strong></a></p> <p><strong>Editorial Director:</strong></p> <p>Leonid Sokolinsky, South Ural State University, Chelyabinsk, Russia</p> <p><strong><a href="https://superfri.org/index.php/superfri/about/#custom-2">Editorial Board</a></strong></p> <p><strong>Production:</strong> South Ural State University (Chelyabinsk, Russia)</p> <p><strong>ISSN:</strong> 2313-8734 (online), 2409-6008 (print) <strong>DOI:</strong> 10.14529/jsfi</p> <p><strong>Publication Frequency:</strong> 4 issues (print and electronic) per year</p> <p><strong>Current Issue:</strong> <a href="https://superfri.org/index.php/superfri/issue/current">Volume 12, Number 3 (2025)</a> <strong>DOI:</strong> 10.14529/jsfi2503.</p> <p><strong>Abstracting and Indexing:</strong> <a href="https://www.scopus.com/sourceid/21100843325">Scopus</a>, <a href="http://dl.acm.org/citation.cfm?id=J1529">ACM Digital Library</a>, <a href="https://doaj.org/toc/2313-8734" target="_blank" rel="noopener">DOAJ</a>.</p> </td> <td align="center" valign="top"><a href="https://superfri.org/index.php/superfri/issue/current"> <img src="https://superfri.org/public/site/images/porozovas/superfri-2022-1-without-ssn.png" alt="" align="top" /><img src="https://superfri.org/public/site/images/kraevaya/superfri-2025-3-without-issn.png" alt="" width="215" height="301" /></a></td> </tr> <tr> <td align="center" valign="top"><a href="https://www.scopus.com/sourceid/21100843325"> <img style="width: 180px;" src="https://superfri.org/public/site/images/kraevaya/citescore2024-supercomputing-front.png" width="35%" height="100" /> </a> <!--<a title="SCImago Journal &amp; Country Rank" href="https://www.scimagojr.com/journalsearch.php?q=21100843325&amp;tip=sid&amp;clean=0"> <img style="margin-top: 1em; width: 60%;" src="https://www.scimagojr.com/journal_img.php?id=21100843325" alt="SCImago Journal &amp; Country Rank" width="35%" border="0" /> </a>--></td> </tr> <tr> <td colspan="2"><strong><a href="https://superfri.org/index.php/superfri/special-issue-vol13-no1-2026">Special Issue "Supercomputing Challenges in Molecular Modeling in Life and Material Sciences and Astrochemistry"</a></strong></td> </tr> </tbody> </table> <div class="separator"> </div> <!--<div class="separator" style="padding: 1em 0em 1em 0em;"><strong>Special Issue on <a href="https://easychair.org/cfp/CAES2023">Computer Aided Engineering on Supercomputers</a></strong> (VOL 10, NO 4 2023, deadline is 20 November 2023)</div>--> en-US <p>Authors retain copyright and grant the journal right of first publication with the work simultaneously licensed under a <a href="http://www.creativecommons.org/licenses/by-nc/3.0/" target="_new">Creative Commons Attribution-Non Commercial 3.0 License</a> that allows others to share the work with an acknowledgement of the work's authorship and initial publication in this journal.</p> voevodin@parallel.ru (Vladimir Voevodin) mzym@susu.ru (Mikhail Zymbler) Thu, 25 Dec 2025 00:00:00 +0500 OJS 3.3.0.4 http://blogs.law.harvard.edu/tech/rss 60 RuParam: a Russian Parametric Dataset for LLM Evaluation https://superfri.susu.ru/index.php/superfri/article/view/637 <p class="p1">We introduce RuParam, a parametric dataset designed to evaluate the acquisition of Russian by large language models (LLMs). This corpus mirrors the structure of the BLiMP family of datasets by containing minimal pairs of sentences. However, our goal was to expand its scope as much as possible by incorporating diverse phenomena from several domains of Russian grammar. A significant portion of the data originates from the Tests of Russian as a Foreign Language (TORFL); similar sources were not previously used for linguistic evaluation of LLMs. Additionally, this study details experimental findings involving six LLMs. These LLMs, sourced from multiple developers, vary in size and pretraining data, which affects their proficiency in Russian. We investigate how effectively these models handle universal, typological, and Russian-specific grammatical features. Our results indicate that while most of the models demonstrate relatively high performance, they struggle significantly with some of the Russian-specific categories.</p> Pavel V. Grashchenkov, Lada I. Pasko, Regina R. Nasyrova Copyright (c) 2025 Supercomputing Frontiers and Innovations https://superfri.susu.ru/index.php/superfri/article/view/637 Thu, 25 Dec 2025 00:00:00 +0500 Large Language Models versus Native Speakers in Emotional Assessment of Russian Words https://superfri.susu.ru/index.php/superfri/article/view/638 <p class="p1">The paper presents a comparative analysis of emotional evaluation of Russian nouns by large language models and native speakers. Based on the ENRuN (Emotional Norms for Russian Nouns) database, which contains ratings of 1,800 nouns across five basic emotions (happiness, sadness, anger, fear, and disgust), the research compares human assessments with evaluations provided by seven large language models (Llama-3-70B, Qwen 2.5-32B, YandexGPT 5 Lite, RuadaptQwen2.5-7B, RuadaptQwen2.5-32B-Pro-Beta, T-pro, T-lite). Although some models demonstrated relatively high correlation with human assessments, persistent systematic deviations were observed across all tested models. The analysis reveals significant differences in emotional perception during word evaluation: the models demonstrate a tendency to hyperbolise negative emotions and show variability in assessing positive emotions, particularly when analysing words related to sensitive topics (violence, religion, obscene vocabulary). The findings indicate that the closest alignment with human evaluations is achieved when there is a balance between the model’s size and the quality of its language adaptation.</p> Polina V. Iaroshenko, Natalia V. Louckachevitch Copyright (c) 2025 Supercomputing Frontiers and Innovations https://superfri.susu.ru/index.php/superfri/article/view/638 Thu, 25 Dec 2025 00:00:00 +0500 LLM for Semantic Role Labeling of Emotion Predicates in Russian https://superfri.susu.ru/index.php/superfri/article/view/639 <p class="p1">Semantic role labeling (SRL) for morphologically rich languages, such as Russian, faces significant challenges due to complex case marking systems, free word order, and limited annotated resources. These challenges are particularly acute for emotion predicates, which require specialized linguistic expertise to capture distinctions between roles denoting those who feel, causes and objects of feelings. We propose a novel approach that leverages large language models to address SRL for Russian emotion predicates through few-shot in-context learning combined with predicate-specific instructions. Our method was evaluated on a manually annotated dataset of 169 sentences containing six emotion predicate groups extracted from Russian social media texts. We compared three state-of-the-art LLMs (Claude 3.7 Sonnet, GPT-5 Mini, and DeepSeek V3) against a RuELECTRA-based trained sequence labelling baseline using both exact and partial matching criteria. Claude 3.7 achieved the highest performance with 74.85% F1 score on partial matching, substantially outperforming the baseline (22.67%). For general predicates on FrameBank, our adapted method with GPT-5 Mini reached 85.0% F1 compared to the previous state-of-the-art of 80.1%. The LLM-based approach successfully handles complex linguistic phenomena, including syntactic zeros and multi-word arguments, while requiring minimal manually annotated training data. We demonstrate that LLM-based methods can significantly advance SRL for Russian by reducing dependency on large-scale annotated corpora while achieving competitive performance.</p> Ivan V. Smirnov, Daniil S. Larionov, Elena N. Nikitina, Grigory A. Kazachonok Copyright (c) 2025 Supercomputing Frontiers and Innovations https://superfri.susu.ru/index.php/superfri/article/view/639 Thu, 25 Dec 2025 00:00:00 +0500 Document-Level Approach to Extracting Argumentation Structures from the Russian Texts of Scientific Communication https://superfri.susu.ru/index.php/superfri/article/view/640 <p class="p1">The study addresses the problem of automatic extraction of argumentative structures in scientific communication texts in Russian. Such texts are characterized by a branched logical structure, including distant references and interrelations. To address these complexities, recent methodological advances attempt to leverage the text itself as a contextual foundation for extracting connections. This study presents a generative approach for extracting argumentative relations, reframing the prediction task as a problem of generating marked-up text and making it an end-to-end approach, rather than the traditional pipeline. Two Russian-language corpora were used in the experiments: the translated corpus of microtexts ruMTC and the annotated corpus of scientific communication texts ArgNetSC. A comparative analysis was conducted to evaluate the performance of T5 architecture models trained with supervised fine-tuning (SFT) and Large Language Models on various Russian-language datasets. To facilitate the analysis of long texts, a text segmentation method using a sliding window was proposed. The evaluation revealed that the highest performance in argumentative relation extraction was consistently achieved on the corpus of microtexts. Notably, the smaller models fine-tuned using the SFT method and large language models that were prompted to generate marked texts demonstrated comparable performance (F<sub><span class="s1">1 </span></sub><span class="s1">~ </span>0.32–0.37). For larger texts, however, this trend did not persist, as the FRED-T5 model outperformed all other models with F<sub><span class="s1">1 </span></sub><span class="s1">~ </span>0.23 on texts related to the genre of scientific articles.</p> Elena A. Sidorova, Irina R. Akhmadeeva, Daria V. Ilina, Irina S. Kononenko, Alexey S. Sery, Yury A. Zagorulko Copyright (c) 2025 Supercomputing Frontiers and Innovations https://superfri.susu.ru/index.php/superfri/article/view/640 Thu, 25 Dec 2025 00:00:00 +0500 Can LLMs Get to the Roots? Evaluating Russian Morpheme Segmentation Capabilities in Large Language Models https://superfri.susu.ru/index.php/superfri/article/view/642 <p class="p1">Automatic morpheme segmentation, a crucial task for morphologically rich languages like Russian, is persistently hindered by a significant drop in performance on words containing out-of-vocabulary (OOV) roots. This issue affects even state-of-the-art models, such as fine-tuned BERT models. This study investigates the potential of modern Large Language Models (LLMs) to address this challenge, focusing on the specific task of root identification in Russian. We evaluate a diverse set of eight state-of-the-art LLMs, including proprietary and open-weight models, using a prompt-based, few-shot learning approach. The models' performance is benchmarked against strong baselines – a fine-tuned RuRoberta model and a CNN ensemble – on a 500-word test set. Our results demonstrate that one model, Gemini 2.5 Pro, surpasses both baselines by approximately 5 percentage points in root identification accuracy. An examination of the model's reasoning capabilities shows that while it can produce logically sound, etymologically-informed analyses, it is also highly prone to factual hallucinations. This work highlights that while LLMs show significant promise in overcoming the OOV root problem, the inconsistency of their reasoning presents a significant obstacle to their direct application, underscoring the need for further research into improving their factuality and consistency.</p> Dmitry A. Morozov, Anna V. Glazkova, Boris L. Iomdin Copyright (c) 2025 Supercomputing Frontiers and Innovations https://superfri.susu.ru/index.php/superfri/article/view/642 Thu, 25 Dec 2025 00:00:00 +0500 RuBookSum: Dataset for Russian Literature Abstractive Summarization https://superfri.susu.ru/index.php/superfri/article/view/643 <p class="p1">The majority of existing Russian document summarization datasets focus on short-form source documents which does not require complex causal analysis or coreference resolutions. Furthermore, processing longer multi-page texts poses a serious challenge to current generation of language models as the limited context window complicates response generation by demanding additional task partitioning. To lay the groundwork for future research of the problem, we introduce RuBookSum, an abstractive summarization dataset for Russian long-form narrative summarization. Our dataset covers documents from various literature domains, including fiction, classic, children books and popular science, and includes high-quality human-written summaries. To establish a baseline, we evaluate popular open-source large language models and provide comprehensive analysis on their performance. Additionally, we propose optimized algorithms for long-document summarization, which enable up to 300% summary generation speed up without significant drops in quality.</p> Denis A. Grigoriev, Daniil V. Khudiakov, Daniil I. Chernyshev Copyright (c) 2025 Supercomputing Frontiers and Innovations https://superfri.susu.ru/index.php/superfri/article/view/643 Thu, 25 Dec 2025 00:00:00 +0500 Do Open Large Language Models Know What, Where, and When? A Case Study with Quiz-Style Questions https://superfri.susu.ru/index.php/superfri/article/view/647 <p class="p1">Large language models (LLMs) are increasingly tested on reasoning-intensive benchmarks, yet their performance on complex quiz-style tasks remains underexplored. In this paper we evaluate modern open-source LLMs on the Russian intellectual game What? Where? When?, a challenging format requiring fact recall, associative reasoning, and interpretation of hidden clues. We introduce a new dataset of 2600 questions (2018–2025), enriched with empirical human team success rates and annotated with structural and thematic clusters. We benchmark 14 recent open models accessible via API using both automatic metrics (Exact Match, BLEU, ROUGE) and an LLM-as-a-Judge framework. The best system, Qwen3-235B-A22B-Thinking, achieved 32.4% accuracy, but still lagging behind the average human team success rate (45.8%). Large-scale reasoning-enabled models consistently outperformed non-reasoning or smaller counterparts, particularly in domains such as technology, ancient world, psychology, and nature. However, omission, wordplay, and proper-name questions remained difficult across all systems. Comparison with CheGeKa (MERA leaderboard) shows that our dataset is substantially harder: while leading proprietary and open models reach EM of 0.534–0.645 and 0.442 on CheGeKa, respectively, the strongest model in our benchmark achieves only 0.255 EM. Correlation analysis indicates that human and model perceptions of difficulty only weakly align, suggesting different problem-solving strategies. Qualitative case studies further show that models excel more in fact recall than in reconstructing hidden logic. Our findings highlight both the progress of open LLMs and their current limitations in quiz-style reasoning. The new dataset offers a complementary and more challenging benchmark for Russian-language evaluation.</p> Anna V. Kuznetsova, Viktor A. Byzov, Ilias V. Aslanov, Evgeny V. Kotelnikov Copyright (c) 2025 Supercomputing Frontiers and Innovations https://superfri.susu.ru/index.php/superfri/article/view/647 Thu, 25 Dec 2025 00:00:00 +0500 One-Shot Prompting for Russian Dependency Parsing https://superfri.susu.ru/index.php/superfri/article/view/648 <p class="p1">This study investigates the application of Large Language Models (LLMs) for dependency parsing of Russian sentences. We evaluated several models (including Qwen, RuAdapt, YandexGPT, T-pro, T-lite, and Llama) in a one-shot mode across multiple Russian treebanks: SynTagRus, GSD, PUD, Poetry, and Taiga. Among the models tested, Llama70 achieved the highest scores in both UAS and LAS. Furthermore, we observed a general trend where larger models tended to perform better. Our analysis also revealed that parsing quality for Qwen4 and RuAdapt4 on the Taiga treebank was notably sensitive to prompt design. However, the results from all LLMs remained lower than those obtained from classical neural parsers. A key challenge encountered by many models was a difference between generated token sets and gold token sets, which was observed in a considerable portion of each treebank. Additionally, the T-pro and T-lite models produced a significant number of extra lines. The implementation for this study is publicly available at https://github.com/Derinhelm/llm_parsing/tree/main.</p> Elena D. Shamaeva, Mikhail M. Tikhomirov, Natalia V. Loukachevitch Copyright (c) 2025 Supercomputing Frontiers and Innovations https://superfri.susu.ru/index.php/superfri/article/view/648 Thu, 25 Dec 2025 00:00:00 +0500 Aspect-Based Sentiment Analysis Using Large Language Models on Museum Visitor Reviews https://superfri.susu.ru/index.php/superfri/article/view/649 <p class="p1">Museum reviews provide rich insight into visitor preferences and can drive useful change within institutions, yet they have attracted little attention in sentiment research owing to limited commercial interest and the multi-thematic nature of reviews. In this study we analysed over 12 000 reviews in Russian for 15 museum sites collected from nine different platforms. Methodologically, we first evaluated traditional approaches: a lexicon-based method utilising sentiment dictionaries and a neural network approach leveraging open-source pre-trained models such as RuBERT. While such methods can be applied to document-level sentiment analysis, where the text is labelled simply as positive or negative, they cannot uncover the specific topics that give rise to these sentiments. Finally, we implemented large language models (LLMs) for aspect-based sentiment analysis to discover positive and negative aspects visitors mention. Our system uses a two-step pipeline that initially extracts positive and negative keywords about each museum site and subsequently categorises these keywords into 14 predetermined categories, enabling the reader to effortlessly discover strong points and areas for improvement. Results include 15 csv tables of positive and negative keywords and 15 year-wise text reports for all objects. While some LLM hallucinations were observed, the outputs were largely realistic. We conclude that LLMs are well suited to this task and offer substantial scope for future research and practical applications in museum evaluation and service improvement.</p> Anastasia V. Kolmogorova, Elizaveta R. Kulikova, Vladislav V. Lobanov Copyright (c) 2025 Supercomputing Frontiers and Innovations https://superfri.susu.ru/index.php/superfri/article/view/649 Thu, 25 Dec 2025 00:00:00 +0500 Neurosymbolic Approach to Processing of Educational Texts for Educational Standard Compliance Analysis https://superfri.susu.ru/index.php/superfri/article/view/650 <p class="p1">This article presents a neurosymbolic approach for analyzing the alignment between textbook content and educational standards. The study addresses the problem of assessing terminological coherence by evaluating a corpus of textbooks against the Russian Federal State Educational Standard. We employ a hybrid methodology combining classical symbolic NLP methods for topic modeling (keyword extraction and term alignment) with qualitative analysis and use of modern large language models for items not found algorithmically. The experimental results on a corpus of 5 textbooks on Physics for the 7th grade and corresponding educational standard indicate a mean coverage of standard topics of 71% across all textbooks with use of the symbolic methods. Application of large language model (ChatGPT 5) for the qualitative analysis recovered 51% keywords initially missed by the abovementioned methods. The findings are relevant for researchers in educational linguistics, computational linguistics, curriculum developers, and textbook authors. The proposed pipeline offers a scalable tool for automating analysis of educational content compliance, reducing the workload for manual expert assessment. This work contributes to the development of AI-assisted methodologies in educational standard alignment and textbook quality control.</p> Nikolai A. Prokopyev, Marina I. Solnyshkina, Valery D. Solovyev Copyright (c) 2025 Supercomputing Frontiers and Innovations https://superfri.susu.ru/index.php/superfri/article/view/650 Thu, 25 Dec 2025 00:00:00 +0500