Shashank Jere;Lizhong Zheng;Ummay Sumaya Khan;Lingjia Liu
{"title":"Reservoir Computing in Wireless Receive Processing: An Information Theoretic Perspective","authors":"Shashank Jere;Lizhong Zheng;Ummay Sumaya Khan;Lingjia Liu","doi":"10.1109/JSAIT.2026.3683606","DOIUrl":"https://doi.org/10.1109/JSAIT.2026.3683606","url":null,"abstract":"Deep learning is playing a transformational role in the physical layer of wireless communications. While demonstrating remarkable empirical performance in applications such as MIMO detection and receive processing, the fundamental reasons underlying this superlative performance remain insufficiently understood. In this work, we advance the development of Explainable AI (xAI) within the physical layer by grounding the analysis of deep learning methods in estimation-theoretic and information-theoretic principles. Specifically, we focus on wireless receive processing, namely channel equalization through reservoir computing (RC)—a recurrent neural network (RNN) framework that has shown strong empirical receive processing performance in both Orthogonal Frequency Division Multiplexing (OFDM) and Orthogonal Time Frequency Space (OTFS) systems. Within RC, we focus on the echo state network (ESN) and its extension, the windowed ESN (WESN). Building on our prior work that establishes the signal processing foundations of the ESN in receive processing, we conduct an exact analytical characterization of the equalization performance of the ESN and WESN and compare them against the optimum Wiener equalizer in a wireless fading channel with inter-symbol interference (ISI). The validity of the theoretical results is established through extensive numerical evaluations.","PeriodicalId":73295,"journal":{"name":"IEEE journal on selected areas in information theory","volume":"7 ","pages":"146-160"},"PeriodicalIF":2.2,"publicationDate":"2026-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147796122","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"In-Memory Bit Error Rate Estimation Using Syndromes of LDPC Codes","authors":"Yotam Gershon;Yuval Cassuto","doi":"10.1109/JSAIT.2026.3686020","DOIUrl":"https://doi.org/10.1109/JSAIT.2026.3686020","url":null,"abstract":"Modern AI systems entail steep energy costs due to massive-scale computations and data transfers; offloading parts of the computations to be performed in-memory holds great potential for reducing both. This paper studies a new architecture proposed for reliable in-memory computations. Its main component is a coding scheme that is designed for both in-memory error-rate estimation/detection and outside-of-memory error correction. Estimation and/or detection are used to decide when the error rate exceeds the tolerance of the computation, at which point error correction is invoked. The coding scheme is based on a nested bilayer LDPC construction, where in particular, the first layer comprises degree-1 variable nodes guaranteeing accurate bit-error rate (BER) estimation and detection. Towards that, we derive a closed-form maximum-likelihood BER estimator for irregular codes, and a gapped hypothesis testing framework for deciding when to decode given some prescribed error-rate tolerance. The performance analysis of the derived estimator includes a closed-form mean-squared-error expression with explicit dependence on the check-degree distribution. For the hypothesis testing the analysis shows the dependence of detection performance on the same degree distribution. Both results reveal an advantage of check-regular codes that minimize dominant error terms among codes with a given average check degree.","PeriodicalId":73295,"journal":{"name":"IEEE journal on selected areas in information theory","volume":"7 ","pages":"161-174"},"PeriodicalIF":2.2,"publicationDate":"2026-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147828787","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Stephen Creagh;Valon Blakaj;Kangyu Zhao;Gabriele Gradoni
{"title":"Electromagnetic Information Theory in Phase Space","authors":"Stephen Creagh;Valon Blakaj;Kangyu Zhao;Gabriele Gradoni","doi":"10.1109/JSAIT.2025.3628112","DOIUrl":"https://doi.org/10.1109/JSAIT.2025.3628112","url":null,"abstract":"An approach to characterising operator spectra using a ray-dynamical phase space, originating from treatments of quantum mechanics, is adapted to calculate degrees of freedom and channel capacities of wireless communication between surfaces. The method is grounded on propagation of correlation functions and exploits the outputs of Eulerian ray-tracing algorithms. It presents results using a signal-to-noise ratio expressed as a function of phase space coordinates, resolving it in terms of direction as well as position. The ability of the phase-space representation to capture the spatial-angular dynamics of propagation makes the methodology suitable for advanced studies of electromagnetic signal and information theory. Examples are offered for flat as well as curved surfaces, communicating in free-space and in confined propagation environments.","PeriodicalId":73295,"journal":{"name":"IEEE journal on selected areas in information theory","volume":"6 ","pages":"446-457"},"PeriodicalIF":2.2,"publicationDate":"2025-11-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145560793","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"On Algebraic Designing of DNA Codes With Biological and Combinatorial Constraints","authors":"Krishna Gopal Benerjee;Adrish Banerjee","doi":"10.1109/JSAIT.2025.3619053","DOIUrl":"https://doi.org/10.1109/JSAIT.2025.3619053","url":null,"abstract":"This paper presents constructions of DNA codes that satisfy biological and combinatorial constraints for DNA-based data storage systems. We introduce an algorithm that generates DNA blocks containing sequences that meet the required constraints for DNA codes. The constructed DNA sequences satisfy biological constraints: balanced GC-content, avoidance of secondary structures, and prevention of homopolymer runs. These sequences simultaneously satisfy combinatorial constraints that ensure differences among DNA sequences and their reverse and reverse-complement sequences. The DNA codes incorporate error correction through minimum Hamming distance requirements. We establish a bijective mapping between algebraic structures and DNA sequences, providing construction of DNA codes with specified characteristics. Using this framework, we construct DNA codes based on error-correcting codes, including Simplex and Reed-Muller codes. These constructions ensure DNA sequences avoid secondary structures and homopolymer runs exceeding length three, which cause errors in DNA storage systems. Concatenated sequences maintain these properties. The codes achieve non-vanishing code rates and minimum Hamming distances for large sequence lengths, demonstrating viability for DNA-based data storage systems.","PeriodicalId":73295,"journal":{"name":"IEEE journal on selected areas in information theory","volume":"6 ","pages":"432-445"},"PeriodicalIF":2.2,"publicationDate":"2025-10-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145405299","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Jin Sima;Chao Pan;S. Kasra Tabatabaei;Alvaro G. Hernandez;Charles M. Schroeder;Olgica Milenkovic
{"title":"DNA Tails for Molecular Flash Memory","authors":"Jin Sima;Chao Pan;S. Kasra Tabatabaei;Alvaro G. Hernandez;Charles M. Schroeder;Olgica Milenkovic","doi":"10.1109/JSAIT.2025.3616940","DOIUrl":"https://doi.org/10.1109/JSAIT.2025.3616940","url":null,"abstract":"DNA-based data storage systems face practical challenges due to the high cost of DNA synthesis. A strategy to address the problem entails encoding data via topological modifications of the DNA sugar-phosphate backbone. The DNA Punchcards system, which introduces nicks (cuts) in the DNA backbone, encodes only one bit per nicking site, limiting density. We propose DNA Tails, a storage paradigm that encodes nonbinary symbols at nicking sites by growing enzymatically synthesized single-stranded DNA of varied lengths. The average tail lengths encode multiple information bits and are controlled via a staggered nicking-tail extension process. We demonstrate the feasibility of this encoding approach experimentally and identify common sources of errors, such as calibration errors and stumped tail growth errors. To mitigate calibration errors, we use rank modulation proposed for flash memory. To correct stumped tail growth errors, we introduce a new family of rank modulation codes that can correct “stuck-at” errors. Our analytical results include constructions for order-optimal-redundancy permutation codes and accompanying encoding and decoding algorithms.","PeriodicalId":73295,"journal":{"name":"IEEE journal on selected areas in information theory","volume":"6 ","pages":"458-469"},"PeriodicalIF":2.2,"publicationDate":"2025-10-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145560794","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Coding Methods for String Reconstruction From Erroneous Prefix-Suffix Compositions","authors":"Zitan Chen","doi":"10.1109/JSAIT.2025.3617251","DOIUrl":"https://doi.org/10.1109/JSAIT.2025.3617251","url":null,"abstract":"The number of zeros and the number of ones in a binary string are referred to as the composition of the string, and the prefix-suffix compositions of a string are a multiset formed by the compositions of the prefixes and suffixes of all possible lengths of the string. In this work, we present binary codes of length <inline-formula> <tex-math>$n$ </tex-math></inline-formula> in which every codeword can be efficiently reconstructed from its erroneous prefix-suffix compositions with at most <inline-formula> <tex-math>$t$ </tex-math></inline-formula> composition errors. All our constructions have decoding complexity polynomial in <inline-formula> <tex-math>$n$ </tex-math></inline-formula> and the best of our constructions has constant rate and can correct <inline-formula> <tex-math>$t=Theta (n)$ </tex-math></inline-formula> errors. As a comparison, no prior constructions can afford to efficiently correct <inline-formula> <tex-math>$t=Theta (n)$ </tex-math></inline-formula> arbitrary composition errors. Additionally, we propose a method of encoding <inline-formula> <tex-math>$h$ </tex-math></inline-formula> arbitrary strings of the same length so that they can be reconstructed from the multiset union of their error-free prefix-suffix compositions, at the expense of <inline-formula> <tex-math>$h$ </tex-math></inline-formula>-fold coding overhead. In contrast, existing methods can only recover <inline-formula> <tex-math>$h$ </tex-math></inline-formula> distinct strings, albeit with code rate asymptotically equal to <inline-formula> <tex-math>$1/h$ </tex-math></inline-formula>. Building on the top of the proposed method, we also present a coding scheme that enables efficient recovery of <inline-formula> <tex-math>$h$ </tex-math></inline-formula> strings from their erroneous prefix-suffix compositions with <inline-formula> <tex-math>$t=Theta (n)$ </tex-math></inline-formula> errors.","PeriodicalId":73295,"journal":{"name":"IEEE journal on selected areas in information theory","volume":"6 ","pages":"394-402"},"PeriodicalIF":2.2,"publicationDate":"2025-10-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145352102","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Optimizing the Decoding Probability and Coverage Ratio of Composite DNA","authors":"Tomer Cohen;Eitan Yaakobi","doi":"10.1109/JSAIT.2025.3613272","DOIUrl":"https://doi.org/10.1109/JSAIT.2025.3613272","url":null,"abstract":"This paper studies two problems that are motivated by the novel recent approach of composite DNA that takes advantage of the DNA synthesis property which generates a huge number of copies for every synthesized strand. Under this paradigm, every composite symbols does not store a single nucleotide but a mixture of the four DNA nucleotides. The first problem studies the expected number of strand reads in order to decode a composite strand or a group of composite strands. In the second problem, our goal is study how to carefully choose a fixed number of mixtures of the DNA nucleotides such that the decoding probability by the maximum likelihood decoder is maximized.","PeriodicalId":73295,"journal":{"name":"IEEE journal on selected areas in information theory","volume":"6 ","pages":"417-431"},"PeriodicalIF":2.2,"publicationDate":"2025-09-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145405334","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Neural Polar Decoders for DNA Data Storage","authors":"Ziv Aharoni;Henry D. Pfister","doi":"10.1109/JSAIT.2025.3610751","DOIUrl":"https://doi.org/10.1109/JSAIT.2025.3610751","url":null,"abstract":"Synchronization errors, arising from both synthesis and sequencing noise, present a fundamental challenge in DNA-based data storage systems. These errors are often modeled as insertion-deletion-substitution (IDS) channels, for which maximum-likelihood decoding is quite computationally expensive. In this work, we propose a data-driven approach based on neural polar decoders (NPDs) to design decoders with reduced complexity for channels with synchronization errors. The proposed architecture enables decoding over IDS channels with reduced complexity <inline-formula> <tex-math>$O(A N log N)$ </tex-math></inline-formula>, where <inline-formula> <tex-math>$A$ </tex-math></inline-formula> is a tunable parameter independent of the channel. NPDs require only sample access to the channel and can be trained without an explicit channel model. Additionally, NPDs provide mutual information (MI) estimates that can be used to optimize input distributions and code design. We demonstrate the effectiveness of NPDs on both synthetic deletion and IDS channels. For deletion channels, we show that NPDs achieve near-optimal decoding performance and accurate MI estimation, with significantly lower complexity than trellis-based decoders. We also provide numerical estimates of the channel capacity for the deletion channel. We extend our evaluation to realistic DNA storage settings, including channels with multiple noisy reads and real-world Nanopore sequencing data. Our results show that NPDs match or surpass the performance of existing methods while using significantly fewer parameters than the state-of-the-art. These findings highlight the promise of NPDs for robust and efficient decoding in DNA data storage systems.","PeriodicalId":73295,"journal":{"name":"IEEE journal on selected areas in information theory","volume":"6 ","pages":"403-416"},"PeriodicalIF":2.2,"publicationDate":"2025-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145405347","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Geno-Weaving: A Framework for Low-Complexity Capacity-Achieving DNA Data Storage","authors":"Hsin-Po Wang;Venkatesan Guruswami","doi":"10.1109/JSAIT.2025.3610643","DOIUrl":"https://doi.org/10.1109/JSAIT.2025.3610643","url":null,"abstract":"As a potential implementation of data storage using DNA molecules, multiple strands of DNA are stored unordered in a liquid container. When the data are needed, an array of DNA readers will sample the strands with replacement, producing a Poisson-distributed number of noisy reads for each strand. The primary challenge here is to design an algorithm that reconstructs data from these unsorted, repetitive, and noisy reads. In this paper, we lay down a capacity-achieving rateless code along each strand to encode its index; we then lay down a capacity-achieving block code at the same position across all strands to protect the data. These codes weave a low-complexity storage scheme that saturates the fundamental upper limit of DNA. This improves upon the previous work of Weinberger and Merhav, which proves said bound and uses high-complexity random codes to saturate the limit. Our scheme also differs from other concatenation-based implementations of DNA data storage in the sense that, instead of decoding the inner codes first and passing the results to the outer code, our decoder alternates between the rateless codes and the block codes.","PeriodicalId":73295,"journal":{"name":"IEEE journal on selected areas in information theory","volume":"6 ","pages":"383-393"},"PeriodicalIF":2.2,"publicationDate":"2025-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145315409","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Tailoring Fault-Tolerance to Quantum Algorithms","authors":"Zhuangzhuang Chen;Narayanan Rengaswamy","doi":"10.1109/JSAIT.2025.3602446","DOIUrl":"https://doi.org/10.1109/JSAIT.2025.3602446","url":null,"abstract":"The standard approach to universal fault-tolerant quantum computing is to develop a general purpose quantum error correction mechanism that can implement a universal set of logical gates fault-tolerantly. Given such a scheme, any quantum algorithm can be realized fault-tolerantly by composing the relevant logical gates from this set. However, we know that quantum computers provide a significant quantum advantage only for specific quantum algorithms. Hence, a universal quantum computer can likely gain from compiling such specific algorithms using tailored quantum error correction schemes. In this work, we take the first steps towards such algorithm-tailored quantum fault-tolerance. We consider Trotter circuits in quantum simulation, which is an important application of quantum computing. We develop a solve-and-stitch algorithm to systematically synthesize physical realizations of Clifford Trotter circuits on the well-known <inline-formula> <tex-math>$[![n,n-2,2]!]$ </tex-math></inline-formula> error-detecting code family. Our analysis shows that this family implements Trotter circuits with essentially optimal depth under reasonable assumptions, thereby serving as an illuminating example of tailored quantum error correction. We achieve fault-tolerance for these circuits using flag gadgets, which add minimal overhead. Importantly, the solve-and-stitch algorithm has the potential to scale beyond this specific example, as illustrated by a generalization to the four-qubit logical Clifford Trotter circuit on the <inline-formula> <tex-math>$[![{ 20,4,2 }]!] $ </tex-math></inline-formula> hypergraph product code, thereby providing a principled approach to tailored fault-tolerance in quantum computing.","PeriodicalId":73295,"journal":{"name":"IEEE journal on selected areas in information theory","volume":"6 ","pages":"311-324"},"PeriodicalIF":2.2,"publicationDate":"2025-08-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144998382","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}