V. A. Zholobov, E. D. Romanenkova, S. A. Egorov, N. A. Gevorgyan, A. A. Zaytsev
{"title":"Universal Representations for Well-Logging Data via Ensembling of Self-Supervised Models","authors":"V. A. Zholobov, E. D. Romanenkova, S. A. Egorov, N. A. Gevorgyan, A. A. Zaytsev","doi":"10.1134/S1064562424602257","DOIUrl":null,"url":null,"abstract":"<p>Time series representation learning is crucial in applications requiring sophisticated data analysis. In some areas, like the Oil and Gas industry, the problem is particularly challenging due to missing values and anomalous samples caused by sensor failures in highly complex manufacturing environments. Self-supervised learning is one of the most popular solutions for obtaining data representation. However, being either generative or contrastive, these methods suffer from the limited applicability of obtained embeddings, – so general usage is more often declared than achieved.</p><p>This study introduces and examines various generative self-supervised architectures for complex industrial time series. Moreover, we propose a new way to ensemble several generative approaches, leveraging the best advantages of each method. The suggested procedure is designed to tackle a wide range of scenarios with missing and multiscale data.</p><p>For numerical experiments, we use various-scale datasets of well logs from diverse oilfields. Evaluation includes change point detection, clustering, and transfer learning, with the last two problems being introduced for the first time. It shows that variational autoencoders excel in clustering, autoregressive models better detect change points, and the proposed ensemble succeeds in both tasks.</p>","PeriodicalId":531,"journal":{"name":"Doklady Mathematics","volume":"110 1 supplement","pages":"S126 - S136"},"PeriodicalIF":0.5000,"publicationDate":"2025-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1134/S1064562424602257.pdf","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Doklady Mathematics","FirstCategoryId":"100","ListUrlMain":"https://link.springer.com/article/10.1134/S1064562424602257","RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"MATHEMATICS","Score":null,"Total":0}
引用次数: 0
Abstract
Time series representation learning is crucial in applications requiring sophisticated data analysis. In some areas, like the Oil and Gas industry, the problem is particularly challenging due to missing values and anomalous samples caused by sensor failures in highly complex manufacturing environments. Self-supervised learning is one of the most popular solutions for obtaining data representation. However, being either generative or contrastive, these methods suffer from the limited applicability of obtained embeddings, – so general usage is more often declared than achieved.
This study introduces and examines various generative self-supervised architectures for complex industrial time series. Moreover, we propose a new way to ensemble several generative approaches, leveraging the best advantages of each method. The suggested procedure is designed to tackle a wide range of scenarios with missing and multiscale data.
For numerical experiments, we use various-scale datasets of well logs from diverse oilfields. Evaluation includes change point detection, clustering, and transfer learning, with the last two problems being introduced for the first time. It shows that variational autoencoders excel in clustering, autoregressive models better detect change points, and the proposed ensemble succeeds in both tasks.
期刊介绍:
Doklady Mathematics is a journal of the Presidium of the Russian Academy of Sciences. It contains English translations of papers published in Doklady Akademii Nauk (Proceedings of the Russian Academy of Sciences), which was founded in 1933 and is published 36 times a year. Doklady Mathematics includes the materials from the following areas: mathematics, mathematical physics, computer science, control theory, and computers. It publishes brief scientific reports on previously unpublished significant new research in mathematics and its applications. The main contributors to the journal are Members of the RAS, Corresponding Members of the RAS, and scientists from the former Soviet Union and other foreign countries. Among the contributors are the outstanding Russian mathematicians.