Редакторы-составители: В.С. Белобородов, А.Р. Гатиатуллин, Р.А. Гильмуллин, Л.Р. Гильмутдинова, А.К. Ковалёв, А.В. Кузнецов, О.В. Попова, Е.В. Тутубалина, А.Ф. Хасьянов, А.А. Шпильман

ОТ СОСТАВИТЕЛЕЙ

Настоящий тематический выпуск журнала «Электронные библиотеки» включает статьи, подготовленные на основе  докладов, представленных на Междисциплинарной научной конференции «ИИ-ЗАМАН». Конференция прошла 17 сентября 2025 года в рамках международного форума «Kazan Digital Week – 2025» и была посвящена фундаментальным и прикладным исследованиям в области искусственного интеллекта.

Основные направления конференции: компьютерное зрение, обработка естественного языка, воплощённый искусственный интеллект и робототехника, применение искусственного интеллекта в научных исследованиях.

Основная цель проведенной конференции — объединение специалистов, исследователей и студентов для обсуждения современных актуальных задач искусственного интеллекта, обмена результатами и опытом, а также содействие междисциплинарному научному диалогу. Организаторами конференции выступили Академия наук Республики Татарстан, Институт искусственного интеллекта AIRI и Университет Иннополис.

 

Published: 04.12.2025

Intelligent Chemist Robot: Towards an Autonomous Laboratory

Musa Shamilevich Adygamov, Anton Olegovich Golub, Emil Rinatovich Saifullin, Timur Rustemovich Gimadiev, Nikita Yurievich Serov Serov
997-1014
Abstract:

This paper describes a hardware and software platform that enables automated chemical syntheses, including the preparation, heating, and mixing of reaction mixtures, as well as post-synthesis dilution sampling and sending for high-performance liquid chromatography (HPLC) analysis, followed by automated processing of the results. A custom Python library, ChemBot, was developed to control individual robotic devices, and a client web server was created to manage the entire system. A web interface was created to view the system status and the progress of syntheses. The performance of the entire platform for performing experiments was tested by performing aldol condensation syntheses, where the ratio of reagents, the catalyst and its amount, the temperature and time of synthesis were varied. Writing custom code to monitor and control the entire system is an important step toward integrating the robotic system with artificial intelligence (AI), which will ultimately enable the transition to an autonomous laboratory, where target molecule prediction and synthesis, experimental execution and analysis, and, if necessary, refinement or modification of the model will be performed automatically, without the need for human intervention.

Design of a Dynamic Expert System for Analyzing the Impact of Climate Effects on Small and Medium Sized Enterprises

Rustam Arifovich Burnashev, Yaroslav Vladislavovich Sergeev
1015-1035
Abstract:

Growing climate instability is creating new challenges and risks for the resilience of small and medium-sized enterprises (SMEs). This article proposes a prototype architecture for a dynamic expert system comprising several key modules: a user interface, a knowledge base, a server application, and a dynamic data update module with real-time APIs. A distinctive feature of the system is the application of Z⁺-number calculus, implemented using the scikit-fuzzy library, which allows for accounting of graded confidence in evaluations. This approach provides more robust and adaptive risk assessments that are sensitive to changes in the quality of input data. Interactive visualization of the results is built upon OpenStreetMap. The system's methodology for self-adaptation of confidence measures based on historical data is described.

Normalization of Text Recognized by Optical Character Recognition using Lightweight LLMS

Vladislav Konstantinovich Vershinin, Ivan Vladimirovich Khodnenko, Sergey Vladimirovich Ivanov
1036-1056
Abstract:

Despite recent progress, Optical Character Recognition (OCR) on historical newspapers still leaves 5–10% character errors. We present a fully automated post-OCR normalization pipeline that combines lightweight 7–8B instruction-tuned LLMs quantized to 4-bit (INT4) with a small set of regex rules. On the BLN600 benchmark (600 pages of 19th-century British newspapers), our best model YandexGPT-5-Instruct Q4 reduces Character Error Rate (CER) from 8.4% to 4.0% (–52.5%) and Word Error Rate (WER) from 20.2% to 6.5% (–67.8%), while raising semantic similarity to 0.962. The system runs on consumer hardware (RTX-4060 Ti, 8 GB VRAM) at about 35 seconds per page and requires no fine-tuning or parallel training data. These results indicate that compact INT4 LLMs are a practical alternative to large checkpoints for post-OCR cleanup of historical documents.

Digital Modeling for Scoping Review in Studying Intergenerational Cultural Congruence

Aisylu Munavirovna Ganieva
1057-1069
Abstract:

The aim of the work is to identify key topics in modern psychological research of cultural congruence using the method of thematic digital modeling of an array of scientific publications.


The modernity and significance of the conducted research is due to the growing importance of cultural congruence in the context of the digital transformation of society, which is changing the ways of socialization and interaction. Modern technologies require rethinking the psychological mechanisms of individual adaptation to the cultural environment, especially in childhood and adolescence. Despite the active study of this phenomenon, there is a noticeable shortage of research on the cultural congruence of adults. The use of digital modeling and artificial intelligence allows us to systematize knowledge and identify the structure of the thematic field with high accuracy. The obtained data opens up the prospect for further study of cultural congruence throughout the entire life cycle.


The thematic field review of cultural congruence research was conducted based on an analysis of digital archives comprising a curated collection of 112 scholarly publications on the topic. The review employed a topic modeling algorithm implemented in the Python programming language and leveraged digital platforms incorporating multimodal neural network–based tools (GigaChat, Qwen, DeepSeek). The data analysis yielded four distinct age groups that reflect the developmental specificity of cultural congruence manifestations: preschoolers, primary school–age children, adolescents, and adults.

Automatic Extraction of Argumentative Relations from Scientific Communication Texts

Yury Alekseevich Zagorulko, Elena Anatolievna Sidorova, Irina Ravilevna Akhmadeeva
1070-1084
Abstract:

The complexity of the problem of extracting argumentative structures is associated with such problems as selecting argumentative segments, predicting long-range connections between non-contact segments, and training on data labeled with a low degree of inter-annotator consistency. In this paper, we consider an approach to extracting argumentative relations from fairly large texts related to scientific communication. A comparative analysis was performed of fine-tuning methods using a pre-trained Longformer-type language model that takes into account long contexts and two methods that take into account annotator discrepancies in argument labeling by using the so-called soft labels obtained by uniformly smoothing labels and averaging expert assessments. The experiments were conducted on four datasets containing positive and negative examples of statement pairs (premise, conclusion) and differing in segmentation methods and average text size. The best results were obtained using the model with averaging expert assessments. At the same time, it is noted that the model using smoothed labels also increases the accuracy of classifiers, but worsens the recall.

Neuro-Symbolic Approach to Augmented Text Generation via Automated Induction of Morphotactic Rules

Marat Vildanovich Isangulov, Alexander Mikhailovich Elizarov, Aygiz Razhapovich Kunafin, Airat Rafizovich Gatiatullin, Nikolai Arkadievich Prokopyev
1085-1102
Abstract:

The work presents a hybrid neuro-symbolic method that combines a large language model (LLM) and a finite-state transducer (FST) to ensure morphological correctness in text generation for agglutinative languages. The system automatically extracts rules from corpus data: for local examples of word forms, the LLM produces sequences of morphological analyses, which are then aggregated and organized into compact descriptions of morphotactic rules (LEXC) and allomorph selection (regex). During generation, the LLM and FST operate jointly: if a token is not recognized by the automaton, the LLM derives a “lemma+tags” pair from the context, and the FST produces the correct surface form. A literary corpus (~1600 sentences) was used as the dataset. For a list of 50 nouns, 250 word forms were extracted. Using the proposed algorithm, the LLM generated 110 context-sensitive regex rules along with LEXC morphotactics, from which an FST was compiled that recognized 170/250 forms (~70%). In an applied machine translation test on a subcorpus of 300 sentences, integrating this FST into the LLM cycle improved quality from BLEU 16.14 / ChrF 45.13 to BLEU 25.71 / ChrF 50.87 without retraining the translator. The approach scales to other parts of speech (verbs, adjectives, etc.) as well as to other agglutinative and low-resource languages, where it can accelerate the development of lexical and grammatical resources.

Measuring Uncertainty in Transformer Circuits with Effective Information Consistency

Anatoly Anatolievich Krasnovsky
1103-1119
Abstract:

Mechanistic interpretability has identified functional subgraphs within large language models (LLMs), known as Transformer Circuits (TCs), that appear to implement specific algorithms. Yet we lack a formal, single-pass way to quantify when an active circuit is behaving coherently and thus likely trustworthy. Building on the author’s prior sheaf‑theoretic formulation of causal emergence (Krasnovsky, 2025), we specialize it to transformer circuits and introduce the single‑pass, dimensionless Effective‑Information Consistency Score (EICS). EICS combines (i) a normalized sheaf inconsistency computed from local Jacobians and activations, with (ii) a Gaussian EI proxy for circuit-level causal emergence derived from the same forward state. The construction is white-box, single-pass, and makes units explicit so that the score is dimensionless. We further provide practical guidance on score interpretation, computational overhead (with fast and exact modes), and a toy sanity-check analysis.

Abstractive Summarization for Trade News Analysis Based on a New Domain-Specific Dataset

Daria Andreevna Lyutova, Valentin Andreevich Malykh
1120-1137
Abstract:

We present TradeNewsSum—a corpus for abstractive summarization of international trade news—covering Russian- and English-language publications from domain-specific sources. All summaries are manually prepared following unified guidelines. We conducted experiments with fine-tuning transformer and seq2seq models and performed automatic evaluation using the LLM-as-a-judge scheme. LLaMA 3.1 in instruction-prompting mode achieved the best results, showing high scores across metrics, including factual completeness.

Exploring Post-Training Quantization of Large Language Models with a Focus on Russian Evaluation

Dmitrii Romanovich Poimanov, Mikhail Sergeevich Shutov
1138-1163
Abstract:

The rapid adoption of large language models (LLMs) has made quantization a central technique for enabling efficient deployment under real-world hardware and memory constraints. While English-centric evaluations of low-bit quantization are increasingly available, much less is known about its effects on morphologically rich and resource-diverse languages such as Russian. This gap is particularly important given the recent emergence of high-performing Russian and multilingual LLMs. In this work, we conduct a systematic study of 2-, 3-, and 4-bit post-training quantization (PTQ) for state-of-the-art Russian LLMs across different model scales (4B and 32B). Our experimental setup covers both standard uniform quantization and specialized low-bit formats, as well as lightweight finetuning for recovery in the most extreme 2-bit setting. Our findings highlight several important trends: (i) the tolerance of Russian LLMs to quantization differs across model families and scales; (ii) 4-bit quantization is generally robust, especially when advanced formats are used; (iii) 3-bit models expose sensitivity to calibration data and scaling strategies; and (iv) 2-bit models, while severely degraded under naive PTQ, can be partially restored through short finetuning. Empirical results show that the model's domain must be considered when using different quantization techniques.

Hiding in Meaning: Semantic Encoding for Generative Text Steganography

Oleg Yurievich Rogov, Dmitrii Evgenievich Indenbom, Dmitrii Sergeevich Korzh, Darya Valeryaevna Pugacheva, Vsevolod Alexandrovich Voronov, Elena Viktorovna Tutubalina
1165-1185
Abstract:

We propose a novel framework for steganographic text generation that hides binary messages within semantically coherent natural language using latent-space conditioning of large language models (LLMs). Secret messages are first encoded into continuous vectors via a learned binary-to-latent mapping, which is used to guide text generation through prefix tuning. Unlike prior token-level or syntactic steganography, our method avoids explicit word manipulation and instead operates entirely within the latent semantic space, enabling more fluent and less detectable outputs. On the receiver side, the latent representation is recovered from the generated text and decoded back into the original message. As a key theoretical contribution, we provide a robustness guarantee: if the recovered latent vector lies within a bounded distance of the original, exact message reconstruction is ensured, with the bound determined by the decoder’s Lipschitz continuity and the minimum logit margin. This formal result offers a principled view of the reliability–capacity trade-off in latent steganographic systems. Empirical evaluation on both synthetic data and real-world domains such as Amazon reviews shows that our method achieves high message recovery accuracy (above 91%), strong text fluency and competitive capacity up to 6 bits per sentence element while maintaining resilience against neural steganalysis. These findings demonstrate that latent conditioned generation offers a secure and practical pathway for embedding information in modern LLMs.

Conditional Electrocardiogram Generation using Hierarchical Variation-al Autoencoders

Ivan Anatolevich Sviridov, Konstantin Sergeevich Egorov
1186-1206
Abstract:

Cardiovascular diseases remain the leading cause of mortality, and automated electrocardiogram (ECG) analysis can ease clinical workloads but is limited by scarce and imbalanced data. Synthetic ECG can mitigate these issues, and while most methods use Generative Adversarial Networks (GANs), recent work show variational autoencoders (VAEs) perform comparably. We introduce cNVAE-ECG, a conditional Nouveau VAE (NVAE) that generates high-resolution, 12-lead, 10-second ECGs with multiple pathologies. Leveraging a compact channel-generation scheme and class embeddings for multi-label conditioning, cNVAE-ECG improves downstream binary and multi-label classification, achieving up to a 2% AUROC gain in transfer learning over GAN-based models.

Where Do the Best Features Lie? A Layer-Wise Analysis of Frozen Encoders for Efficient Endoscopic Image Classification

Ahmad Taha, Rustam A. Lukmanov
1207-1229
Abstract:

In our quest to advance medical AI, we demonstrate that a pre-trained and frozen Vision Transformer paired with a linear classifier can achieve highly competitive performance in endoscopic image classification. Our central contribution is a systematic, layer-wise analysis that identifies the source of the most powerful features, challenging the common heuristic of using only the final layer. We uncover a distinct "peak-before-the-end" phenomenon, where a late-intermediate layer offers a more generalizable representation for the downstream medical task. On the Kvasir and HyperKvasir benchmarks, our parameter-light approach not only achieves excellent accuracy but also drastically reduces computational overhead. This work provides a practical roadmap for efficiently leveraging the power of general foundation models in clinical environments.

Verified Explainability Core: a GD-ANFIS/SHAP Hybrid Architecture for XAI 2.0

Yuri Vladislavovich Trofimov, Alexander Dmitrievich Lebedev, Andrei Sergeevich Ilin, Alexey Nikolaevich Averkin
1230-1252
Abstract:

This paper proposes a hybrid Explainable AI architecture that fuses a fully differentiable neuro-fuzzy GD-ANFIS model with the post-hoc SHAP method. The integration is designed to meet XAI 2.0 principles, which call for explanations that are transparent, verifiable, and adaptable at the same time. GD-ANFIS produces human-readable Takagi-Sugeno rules, ensuring structural interpretability, whereas SHAP delivers quantitative feature contributions derived from Shapley theory. To merge these layers, we introduce a comparative-audit mechanism that automatically matches the sets of key features identified by both methods, checks whether the directions of influence coincide, and assesses the consistency between SHAP numerical scores and GD-ANFIS linguistic rules. Such dual-loop on global soil-subsidence mapping, and RMSE 2.30 and 2.36 on Boston Housing and surface-water-quality monitoring respectively, all with full interpretability preserved. In every case, top-feature overlap between the two explanation layers exceeded 60%, demonstrating strong agreement between structural and numerical interpretations. The proposed architecture therefore offers a practical foundation for responsible XAI 2.0 deployment in critical domains ranging from medicine and ecology to geoinformation systems and finance.

AI in Cancer Prevention: a Retrospective Study

Petr Aleksandrovich Philonenko, Vladimir Nikolaevich Kokh, Pavel Dmitrievich Blinov
1253-1266
Abstract:

This study investigates the feasibility of effectively solving population-scale cancer screening problems using artificial intelligence (AI) methods that predict malignant neoplasm risk based on minimal electronic health record (EHR) data – medical diagnosis and service codes. To address the formulated problem, we considered a broad spectrum of modern approaches, including classical machine learning methods, survival analysis, deep learning, and large language models (LLMs). Numerical experiments demonstrated that gradient boosting using survival analysis models as additional predictors possesses the best ability to rank patients by cancer risk level, enabling consideration of both population-level and individual risk factors for malignant neoplasms. Predictors constructed from EHR data include demographic characteristics, healthcare utilization patterns, and clinical markers. This solution was tested in retrospective experiments under the supervision of specialized oncologists. In the retrospective experiment involving more than 1.9 million patients, we established that the risk group captures up to 5.4 times more patients with cancer at the same level of medical examinations. The investigated method represents a scalable solution using exclusively diagnosis and service codes, requiring no specialized infrastructure and integrable into oncological vigilance processes, making it applicable for population-scale cancer screening.

Stylometric Analysis in the Task of Searching for Borrowings of Texts in the Tatar Language

Izida Zufarovna Khayaleeva, Mikhail Mikhailovich Abramskiy
1267-1278
Abstract:

This article discusses the use of stylometric analysis in searching for borrowings of text in the Tatar language. Relevant tools have been developed, utilizing machine learn-ing algorithms, including clustering (k-means method), classification (random forest method, support vector machine method, naive Bayes classifier), and a hybrid approach (FastText model + logistic regression). Special attention is paid to the adaptation of lin-guistic metrics for the Tatar language.