S. Ignatiev, V. Egiazarian, R. Rakhimov, E. Burnaev
{"title":"Deforming Implicit Neural Representation Generative Adversarial Network for Unsupervised Appearence Editing","authors":"S. Ignatiev, V. Egiazarian, R. Rakhimov, E. Burnaev","doi":"10.1134/S1064562424602373","DOIUrl":"10.1134/S1064562424602373","url":null,"abstract":"<p>In this work, we present a new deep generative model for disentangling image shape from its appearance through differentiable warping. We propose to use implicit neural representations for modeling the deformation field and show that coordinate-based representations hold the necessary inductive bias. Unlike the previous warping-based approaches, which tend to model only local and small-scale displacements, our method is able to learn complex deformations and is not restricted to reversible mappings. We study the convergence of warping-based generative models and find that the high-frequency nature of the textures leads to shattered learning gradients, slow convergence, and suboptimal solutions. To cope with this problem, we propose to use invertible blurring, which smooths the gradients and leads to improved results. As a way to further facilitate the convergence of warping, we train the deformation module jointly as a vanilla GAN generator to guide the learning process in a self-distillation manner. Our complete pipeline shows decent results on the LSUN churches dataset. Finally, we demonstrate various applications of our model, like composable texture editing, controllable deformation editing, and keypoint detection.</p>","PeriodicalId":531,"journal":{"name":"Doklady Mathematics","volume":"110 1 supplement","pages":"S299 - S311"},"PeriodicalIF":0.5,"publicationDate":"2025-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1134/S1064562424602373.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143676211","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
V. A. Vasilev, V. S. Arkhipkin, J. D. Agafonova, T. V. Nikulina, E. O. Mironova, A. A. Shichanina, N. A. Gerasimenko, M. A. Shoytov, D. V. Dimitrov
{"title":"CRAFT: Cultural Russian-Oriented Dataset Adaptation for Focused Text-to-Image Generation","authors":"V. A. Vasilev, V. S. Arkhipkin, J. D. Agafonova, T. V. Nikulina, E. O. Mironova, A. A. Shichanina, N. A. Gerasimenko, M. A. Shoytov, D. V. Dimitrov","doi":"10.1134/S1064562424602324","DOIUrl":"10.1134/S1064562424602324","url":null,"abstract":"<p>Despite the fact that popular text-to-image generation models cope well with international and general cultural queries, they have a significant knowledge gap regarding individual cultures. This is due to the content of existing large training datasets collected on the Internet, which are predominantly based on Western European or American popular culture. Meanwhile, the lack of cultural adaptation of the model can lead to incorrect results, a decrease in the generation quality, and the spread of stereotypes and offensive content. In an effort to address this issue, we examine the concept of cultural code and recognize the critical importance of its understanding by modern image generation models, an issue that has not been sufficiently addressed in the research community to date. We propose the methodology for collecting and processing the data necessary to form a dataset based on the cultural code, in particular the Russian one. We explore how the collected data affects the quality of generations in the national domain and analyze the effectiveness of our approach using the Kandinsky 3.1 text-to-image model. Human evaluation results demonstrate an increase in the level of awareness of Russian culture in the model.</p>","PeriodicalId":531,"journal":{"name":"Doklady Mathematics","volume":"110 1 supplement","pages":"S137 - S150"},"PeriodicalIF":0.5,"publicationDate":"2025-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143676299","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"SwiftDepth++: An Efficient and Lightweight Model for Accurate Depth Estimation","authors":"Y. Dayoub, I. Makarov","doi":"10.1134/S1064562424602038","DOIUrl":"10.1134/S1064562424602038","url":null,"abstract":"<p>Depth estimation is a crucial task across various domains, but the high cost of collecting labeled depth data has led to growing interest in self-supervised monocular depth estimation methods. In this paper, we introduce SwiftDepth++, a lightweight depth estimation model that delivers competitive results while maintaining a low computational budget. The core innovation of SwiftDepth++ lies in its novel depth decoder, which enhances efficiency by rapidly compressing features while preserving essential information. Additionally, we incorporate a teacher-student knowledge distillation framework that guides the student model in refining its predictions. We evaluate SwiftDepth++ on the KITTI and NYU datasets, where it achieves an absolute relative error (Abs_rel) of 10.2% on the KITTI dataset and 22% on the NYU dataset without fine-tuning, all with approximately 6 million parameters. These results demonstrate that SwiftDepth++ not only meets the demands of modern depth estimation tasks but also significantly reduces computational complexity, making it a practical choice for real-world applications.</p>","PeriodicalId":531,"journal":{"name":"Doklady Mathematics","volume":"110 1 supplement","pages":"S162 - S171"},"PeriodicalIF":0.5,"publicationDate":"2025-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1134/S1064562424602038.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143676272","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Hiding Backdoors within Event Sequence Data via Poisoning Attacks","authors":"A. Ermilova, E. Kovtun, D. Berestnev, A. Zaytsev","doi":"10.1134/S1064562424602221","DOIUrl":"10.1134/S1064562424602221","url":null,"abstract":"<p>Deep learning’s emerging role in the financial sector’s decision-making introduces risks of adversarial attacks. A specific threat is a poisoning attack that modifies the training sample to develop a backdoor that persists during model usage. However, data cleaning procedures and routine model checks are easy-to-implement actions that prevent the usage of poisoning attacks. The problem is even more challenging for event sequence models, for which it is hard to design an attack due to the discrete nature of the data. We start with a general investigation of the possibility of poisoning for event sequence models. Then, we propose a concealed poisoning attack that can bypass natural banks’ defences. The empirical investigation shows that the developed poisoned model trained on contaminated data passes the check procedure, being similar to a clean model, and simultaneously contains a simple to-implement backdoor.</p>","PeriodicalId":531,"journal":{"name":"Doklady Mathematics","volume":"110 1 supplement","pages":"S288 - S298"},"PeriodicalIF":0.5,"publicationDate":"2025-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1134/S1064562424602221.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143676207","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Rethinking Graph Classification Problem in Presence of Isomorphism","authors":"S. Ivanov, S. Sviridov, E. Burnaev","doi":"10.1134/S1064562424602385","DOIUrl":"10.1134/S1064562424602385","url":null,"abstract":"<p>There is an increasing interest in developing new models for graph classification problem that serves as a common benchmark for evaluation and comparison of GNNs and graph kernels. To ensure a fair comparison of the models several commonly used datasets exist and current assessments and conclusions rely on the validity of these datasets. However, as we show in this paper majority of these datasets contain isomorphic copies of the data points, which can lead to misleading conclusions. For example, the relative ranking of the graph models can change substantially if we remove isomorphic graphs in the test set.</p><p>To mitigate this we present several results. We show that explicitly incorporating the knowledge of isomorphism in the datasets can significantly boost the performance of any graph model. Finally, we re-evaluate commonly used graph models on refined graph datasets and provide recommendations for designing new datasets and metrics for graph classification problem.</p>","PeriodicalId":531,"journal":{"name":"Doklady Mathematics","volume":"110 1 supplement","pages":"S312 - S331"},"PeriodicalIF":0.5,"publicationDate":"2025-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1134/S1064562424602385.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143676389","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
A. Vatolin, N. Gerasimenko, A. Ianina, K. Vorontsov
{"title":"RuSciBench: Open Benchmark for Russian and English Scientific Document Representations","authors":"A. Vatolin, N. Gerasimenko, A. Ianina, K. Vorontsov","doi":"10.1134/S1064562424602191","DOIUrl":"10.1134/S1064562424602191","url":null,"abstract":"<p>Sharing scientific knowledge in the community is an important endeavor. However, most papers are written in English, which makes dissemination of knowledge in countries where English is not spoken by the majority of people harder. Nowadays, machine translation and language models may help to solve this problem, but it is still complicated to train and evaluate models in languages other than English with no or little data in the required language. To address this, we propose the first benchmark for evaluating models on scientific texts in Russian. It consists of papers from Russian electronic library of scientific publications. We also present a set of tasks which can be used to fine-tune various models on our data and provide a detailed comparison between state-of-the-art models on our benchmark.</p>","PeriodicalId":531,"journal":{"name":"Doklady Mathematics","volume":"110 1 supplement","pages":"S251 - S260"},"PeriodicalIF":0.5,"publicationDate":"2025-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1134/S1064562424602191.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143676486","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Multimodal 3D Map Reconstruction for Intelligent Robotcs Using Neural Network-Based Methods","authors":"D. A. Yudin","doi":"10.1134/S1064562424602014","DOIUrl":"10.1134/S1064562424602014","url":null,"abstract":"<p>Methods for constructing multimodal 3D maps are becoming increasingly important for robot navigation systems. In such maps, each 3D point or object contains, in addition to color and semantic category information, compressed vector representations of a text description or sound. This allows solving problems of moving to objects based on natural language queries, even those that do not explicitly mention the object. This article proposes an original taxonomy of methods that allow constructing multimodal 3D maps using neural network methods. It is shown that sparse methods that use a scene representation in the form of an object graph and large language models to find an answer to spatial and semantic queries demonstrate the most promising results on existing open benchmarks. Based on the analysis, recommendations are formulated for choosing certain methods for solving practical problems of intelligent robotics.</p>","PeriodicalId":531,"journal":{"name":"Doklady Mathematics","volume":"110 1 supplement","pages":"S117 - S125"},"PeriodicalIF":0.5,"publicationDate":"2025-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1134/S1064562424602014.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143676296","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Stack More LLM’s: Efficient Detection of Machine-Generated Texts via Perplexity Approximation","authors":"G. M. Gritsai, I. A. Khabutdinov, A. V. Grabovoy","doi":"10.1134/S1064562424602075","DOIUrl":"10.1134/S1064562424602075","url":null,"abstract":"<p>The development of large language models (LLMs) is currently receiving a great amount of interest, but an update of text generation methods should entail a continuous update of methods for detecting machine-generated texts. Earlier, it has been highlighted that values of perplexity and log-probability are able to capture a measure of the difference between artificial and human-written texts. Using this observation, we define a new criterion based on these two values to judge whether a passage is generated from a given LLM. In this paper, we propose a novel efficient method that enables the detection of machine-generated fragments using an approximation of the LLM perplexity value based on pre-collected statistical language models. Approximation lends a hand in achieving high performance and quality metrics also on fragments from weights-closed LLMs. A large number of pre-collected statistical dictionaries results in an increased generalisation ability and the possibility to cover text sequences from the wild. Such approach is easy to update by only adding a new dictionary with latest model text outputs. The presented method has a high performance and achieves quality with an average of 94% recall in detecting generated fragments among texts from various open-source LLMs. In addition, the method is able to perform in milliseconds, which outperforms state-of-the-art models by a factor of thousands.</p>","PeriodicalId":531,"journal":{"name":"Doklady Mathematics","volume":"110 1 supplement","pages":"S203 - S211"},"PeriodicalIF":0.5,"publicationDate":"2025-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1134/S1064562424602075.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143676484","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"An Empirical Scrutinization of Four Crisp Clustering Methods with Four Distance Metrics and One Straightforward Interpretation Rule","authors":"T. A. Alvandyan, S. Shalileh","doi":"10.1134/S1064562424602002","DOIUrl":"10.1134/S1064562424602002","url":null,"abstract":"<p>Clustering has always been in great demand by scientific and industrial communities. However, due to the lack of ground truth, interpreting its obtained results can be debatable. The current research provides an empirical benchmark on the efficiency of three popular and one recently proposed crisp clustering methods. To this end, we extensively analyzed these (four) methods by applying them to nine real-world and 420 synthetic datasets using four different values of <i>p</i> in Minkowski distance. Furthermore, we validated a previously proposed yet not well-known straightforward rule to interpret the recovered clusters. Our computations showed (i) Nesterov gradient descent clustering is the most effective clustering method using our real-world data, while K-Means had edge over it using our synthetic data; (ii) Minkowski distance with <i>p</i> = 1 is the most effective distance function, (iii) the investigated cluster interpretation rule is intuitive and valid.</p>","PeriodicalId":531,"journal":{"name":"Doklady Mathematics","volume":"110 1 supplement","pages":"S236 - S250"},"PeriodicalIF":0.5,"publicationDate":"2025-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1134/S1064562424602002.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143676485","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
V. A. Es’kin, D. V. Davydov, E. D. Egorova, A. O. Malkhanov, M. A. Akhukov, M. E. Smorkalov
{"title":"About Modifications of the Loss Function for the Causal Training of Physics-Informed Neural Networks","authors":"V. A. Es’kin, D. V. Davydov, E. D. Egorova, A. O. Malkhanov, M. A. Akhukov, M. E. Smorkalov","doi":"10.1134/S106456242460194X","DOIUrl":"10.1134/S106456242460194X","url":null,"abstract":"<p>A method is presented that allows to reduce a problem described by differential equations with initial and boundary conditions to a problem described only by differential equations which encapsulate initial and boundary conditions. It becomes possible to represent the loss function for physics-informed neural networks (PINNs) methodology in the form of a single term associated with modified differential equations. Thus eliminating the need to tune the scaling coefficients for the terms of loss function related to boundary and initial conditions. The weighted loss functions respecting causality were modified and new weighted loss functions, based on generalized functions, are derived. Numerical experiments have been carried out for a number of problems, demonstrating the accuracy of the proposed approaches. The neural network architecture was proposed for the Korteweg–De Vries equation, which is more relevant for this problem under consideration, and it demonstrates superior extrapolation of the solution in the space-time domain where training was not performed.</p>","PeriodicalId":531,"journal":{"name":"Doklady Mathematics","volume":"110 1 supplement","pages":"S172 - S192"},"PeriodicalIF":0.5,"publicationDate":"2025-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1134/S106456242460194X.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143676273","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}