{"title":"Linear Complementary Dual Codes Constructed from Reinforcement Learning","authors":"Yansheng Wu, Jin Ma, Shandong Yang","doi":"arxiv-2409.08114","DOIUrl":"https://doi.org/arxiv-2409.08114","url":null,"abstract":"Recently, Linear Complementary Dual (LCD) codes have garnered substantial\u0000interest within coding theory research due to their diverse applications and\u0000favorable attributes. This paper directs its attention to the construction of\u0000binary and ternary LCD codes leveraging curiosity-driven reinforcement learning\u0000(RL). By establishing reward and devising well-reasoned mappings from actions\u0000to states, it aims to facilitate the successful synthesis of binary or ternary\u0000LCD codes. Experimental results indicate that LCD codes constructed using RL\u0000exhibit slightly superior error-correction performance compared to those\u0000conventionally constructed LCD codes and those developed via standard RL\u0000methodologies. The paper introduces novel binary and ternary LCD codes with\u0000enhanced minimum distance bounds. Finally, it showcases how Random Network\u0000Distillation aids agents in exploring beyond local optima, enhancing the\u0000overall performance of the models without compromising convergence.","PeriodicalId":501082,"journal":{"name":"arXiv - MATH - Information Theory","volume":"74 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142217187","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Huquan Kang, Luoyi Fu, Russell J. Funk, Xinbing Wang, Jiaxin Ding, Shiyu Liang, Jianghao Wang, Lei Zhou, Chenghu Zhou
{"title":"Scientific and technological knowledge grows linearly over time","authors":"Huquan Kang, Luoyi Fu, Russell J. Funk, Xinbing Wang, Jiaxin Ding, Shiyu Liang, Jianghao Wang, Lei Zhou, Chenghu Zhou","doi":"arxiv-2409.08349","DOIUrl":"https://doi.org/arxiv-2409.08349","url":null,"abstract":"The past few centuries have witnessed a dramatic growth in scientific and\u0000technological knowledge. However, the nature of that growth - whether\u0000exponential or otherwise - remains controversial, perhaps partly due to the\u0000lack of quantitative characterizations. We evaluated knowledge as a collective\u0000thinking structure, using citation networks as a representation, by examining\u0000extensive datasets that include 213 million publications (1800-2020) and 7.6\u0000million patents (1976-2020). We found that knowledge - which we conceptualize\u0000as the reduction of uncertainty in a knowledge network - grew linearly over\u0000time in naturally formed citation networks that themselves expanded\u0000exponentially. Moreover, our results revealed inflection points in the growth\u0000of knowledge that often corresponded to important developments within fields,\u0000such as major breakthroughs, new paradigms, or the emergence of entirely new\u0000areas of study. Around these inflection points, knowledge may grow rapidly or\u0000exponentially on a local scale, although the overall growth rate remains linear\u0000when viewed globally. Previous studies concluding an exponential growth of\u0000knowledge may have focused primarily on these local bursts of rapid growth\u0000around key developments, leading to the misconception of a global exponential\u0000trend. Our findings help to reconcile the discrepancy between the perceived\u0000exponential growth and the actual linear growth of knowledge by highlighting\u0000the distinction between local and global growth patterns. Overall, our findings\u0000reveal major science development trends for policymaking, showing that\u0000producing knowledge is far more challenging than producing papers.","PeriodicalId":501082,"journal":{"name":"arXiv - MATH - Information Theory","volume":"73 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142268489","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"CROSS: A Contributor-Project Interaction Lifecycle Model for Open Source Software","authors":"Tapajit Dey, Brian Fitzgerald, Sherae Daniel","doi":"arxiv-2409.08267","DOIUrl":"https://doi.org/arxiv-2409.08267","url":null,"abstract":"Despite the widespread adoption of open source software (OSS), its\u0000sustainability remains a critical concern, particularly in light of security\u0000vulnerabilities and the often inadequate end-of-service (EoS) processes for OSS\u0000projects as they decline. Existing models of OSS community participation, like\u0000the Onion model and the episodic contribution model, offer valuable insights\u0000but are fundamentally incompatible and fail to provide a comprehensive picture\u0000of contributor engagement with OSS projects. This paper addresses these gaps by\u0000proposing the CROSS model, a novel contributor-project interaction lifecycle\u0000model for open source, which delineates the various lifecycle stages of\u0000contributor-project interaction along with the driving and retaining forces\u0000pertinent to each stage. By synthesizing existing research on OSS communities,\u0000organizational behavior, and human resource development, it explains a range of\u0000archetypal cases of contributor engagement and highlights research gaps,\u0000especially in EoS/offboarding scenarios. The CROSS model provides a foundation\u0000for understanding and enhancing the sustainability of OSS projects, offering a\u0000robust foundation for future research and practical application.","PeriodicalId":501082,"journal":{"name":"arXiv - MATH - Information Theory","volume":"35 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142217191","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Shi-Yuan Wang, Keerthi S. K. Arumugam, Matthieu R. Bloch
{"title":"Bounds on Covert Capacity in the Sub-Exponential Slotted Asynchronous Regime","authors":"Shi-Yuan Wang, Keerthi S. K. Arumugam, Matthieu R. Bloch","doi":"arxiv-2409.07777","DOIUrl":"https://doi.org/arxiv-2409.07777","url":null,"abstract":"We develop tight bounds for the covert capacity of slotted asynchronous\u0000binary-input Discrete Memoryless Channels (DMCs) and Additive White Gaussian\u0000Noise (AWGN) channels, in which a codeword is transmitted in one of several\u0000slots with known boundaries, where the number of slots is sub-exponential in\u0000the codeword length. Our upper and lower bounds are within a multiplicative\u0000factor of $sqrt{2}$ independent of the channel. This result partially fills a\u0000characterization gap between the covert capacity without asynchronism and the\u0000covert capacity with exponential asynchronism. Our key technical contributions\u0000consist of i) a tight upper bound for the relative entropy characterizing the\u0000effect of asynchronism on the covertness constraint in our achievability proof;\u0000ii) a careful converse analysis to characterize the maximum allowable weight or\u0000power of codewords to meet the covertness constraint. Our results suggest that,\u0000unlike the case without asynchronism, the choice of covertness metric does not\u0000change the covert capacity in the presence of asynchronism.","PeriodicalId":501082,"journal":{"name":"arXiv - MATH - Information Theory","volume":"56 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142217188","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Seyed Mohammad Azimi-Abarghouyi, Leandros Tassiulas
{"title":"Over-the-Air Federated Learning via Weighted Aggregation","authors":"Seyed Mohammad Azimi-Abarghouyi, Leandros Tassiulas","doi":"arxiv-2409.07822","DOIUrl":"https://doi.org/arxiv-2409.07822","url":null,"abstract":"This paper introduces a new federated learning scheme that leverages\u0000over-the-air computation. A novel feature of this scheme is the proposal to\u0000employ adaptive weights during aggregation, a facet treated as predefined in\u0000other over-the-air schemes. This can mitigate the impact of wireless channel\u0000conditions on learning performance, without needing channel state information\u0000at transmitter side (CSIT). We provide a mathematical methodology to derive the\u0000convergence bound for the proposed scheme in the context of computational\u0000heterogeneity and general loss functions, supplemented with design insights.\u0000Accordingly, we propose aggregation cost metrics and efficient algorithms to\u0000find optimized weights for the aggregation. Finally, through numerical\u0000experiments, we validate the effectiveness of the proposed scheme. Even with\u0000the challenges posed by channel conditions and device heterogeneity, the\u0000proposed scheme surpasses other over-the-air strategies by an accuracy\u0000improvement of 15% over the scheme using CSIT and 30% compared to the one\u0000without CSIT.","PeriodicalId":501082,"journal":{"name":"arXiv - MATH - Information Theory","volume":"2 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142227011","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"List-based Optimization of Proximal Decoding for LDPC Codes","authors":"Andreas Tsouchlos, Holger Jäkel, Laurent Schmalen","doi":"arxiv-2409.07278","DOIUrl":"https://doi.org/arxiv-2409.07278","url":null,"abstract":"In this paper, the proximal decoding algorithm is considered within the\u0000context of additive white Gaussian noise (AWGN) channels. An analysis of the\u0000convergence behavior of the algorithm shows that proximal decoding inherently\u0000enters an oscillating behavior of the estimate after a certain number of\u0000iterations. Due to this oscillation, frame errors arising during decoding can\u0000often be attributed to only a few remaining wrongly decoded bit positions. In\u0000this letter, an improvement of the proximal decoding algorithm is proposed by\u0000establishing an additional step, in which these erroneous positions are\u0000attempted to be corrected. We suggest an empirical rule with which the\u0000components most likely needing correction can be determined. Using this insight\u0000and performing a subsequent ``ML-in-the-list'' decoding, a gain of up to 1 dB\u0000is achieved compared to conventional proximal decoding, depending on the\u0000decoder parameters and the code.","PeriodicalId":501082,"journal":{"name":"arXiv - MATH - Information Theory","volume":"20 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142217190","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A High-Performance List Decoding Algorithm for Surface Codes with Erroneous Syndrome","authors":"Jifan Liang, Qianfan Wang, Lvzhou Li, Xiao Ma","doi":"arxiv-2409.06979","DOIUrl":"https://doi.org/arxiv-2409.06979","url":null,"abstract":"Quantum error-correcting codes (QECCs) are necessary for fault-tolerant\u0000quantum computation. Surface codes are a class of topological QECCs that have\u0000attracted significant attention due to their exceptional error-correcting\u0000capabilities and easy implementation. In the decoding process of surface codes,\u0000the syndromes are crucial for error correction, though they are not always\u0000correctly measured. Most of the existing decoding algorithms for surface codes\u0000are not equipped to handle erroneous syndrome information or need additional\u0000measurements to correct syndromes with errors, which implies a potential\u0000increase in inference complexity and decoding latency. In this paper, we\u0000propose a high-performance list decoding algorithm for surface codes with\u0000erroneous syndromes. More specifically, to cope with erroneous syndrome\u0000information, we incorporate syndrome soft information, allowing the syndrome to\u0000be listed as well. To enhance the efficiency of the list decoding algorithm, we\u0000use LCOSD, which can significantly reduce the average list size in classical\u0000error correction compared with the conventional ordered statistics decoding\u0000(OSD). Numerical results demonstrate that our proposed algorithm significantly\u0000improves the decoding performance of surface codes with erroneous syndromes\u0000compared to minimum-weight perfect matching (MWPM) and BP decoders.","PeriodicalId":501082,"journal":{"name":"arXiv - MATH - Information Theory","volume":"59 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142217199","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Statistically Valid Information Bottleneck via Multiple Hypothesis Testing","authors":"Amirmohammad Farzaneh, Osvaldo Simeone","doi":"arxiv-2409.07325","DOIUrl":"https://doi.org/arxiv-2409.07325","url":null,"abstract":"The information bottleneck (IB) problem is a widely studied framework in\u0000machine learning for extracting compressed features that are informative for\u0000downstream tasks. However, current approaches to solving the IB problem rely on\u0000a heuristic tuning of hyperparameters, offering no guarantees that the learned\u0000features satisfy information-theoretic constraints. In this work, we introduce\u0000a statistically valid solution to this problem, referred to as IB via multiple\u0000hypothesis testing (IB-MHT), which ensures that the learned features meet the\u0000IB constraints with high probability, regardless of the size of the available\u0000dataset. The proposed methodology builds on Pareto testing and learn-then-test\u0000(LTT), and it wraps around existing IB solvers to provide statistical\u0000guarantees on the IB constraints. We demonstrate the performance of IB-MHT on\u0000classical and deterministic IB formulations, validating the effectiveness of\u0000IB-MHT in outperforming conventional methods in terms of statistical robustness\u0000and reliability.","PeriodicalId":501082,"journal":{"name":"arXiv - MATH - Information Theory","volume":"4 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142217189","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Decomposition of surprisal: Unified computational model of ERP components in language processing","authors":"Jiaxuan Li, Richard Futrell","doi":"arxiv-2409.06803","DOIUrl":"https://doi.org/arxiv-2409.06803","url":null,"abstract":"The functional interpretation of language-related ERP components has been a\u0000central debate in psycholinguistics for decades. We advance an\u0000information-theoretic model of human language processing in the brain in which\u0000incoming linguistic input is processed at first shallowly and later with more\u0000depth, with these two kinds of information processing corresponding to distinct\u0000electroencephalographic signatures. Formally, we show that the information\u0000content (surprisal) of a word in context can be decomposed into two quantities:\u0000(A) heuristic surprise, which signals shallow processing difficulty for a word,\u0000and corresponds with the N400 signal; and (B) discrepancy signal, which\u0000reflects the discrepancy between shallow and deep interpretations, and\u0000corresponds to the P600 signal. Both of these quantities can be estimated\u0000straightforwardly using modern NLP models. We validate our theory by\u0000successfully simulating ERP patterns elicited by a variety of linguistic\u0000manipulations in previously-reported experimental data from six experiments,\u0000with successful novel qualitative and quantitative predictions. Our theory is\u0000compatible with traditional cognitive theories assuming a `good-enough'\u0000heuristic interpretation stage, but with a precise information-theoretic\u0000formulation. The model provides an information-theoretic model of ERP\u0000components grounded on cognitive processes, and brings us closer to a\u0000fully-specified neuro-computational model of language processing.","PeriodicalId":501082,"journal":{"name":"arXiv - MATH - Information Theory","volume":"33 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142217194","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Typicality, entropy and the generalization of statistical mechanics","authors":"Bernat Corominas-Murtra, Rudolf Hanel, Petr Jizba","doi":"arxiv-2409.06537","DOIUrl":"https://doi.org/arxiv-2409.06537","url":null,"abstract":"When at equilibrium, large-scale systems obey conventional thermodynamics\u0000because they belong to microscopic configurations (or states) that are typical.\u0000Crucially, the typical states usually represent only a small fraction of the\u0000total number of possible states, and yet the characterization of the set of\u0000typical states -- the typical set -- alone is sufficient to describe the\u0000macroscopic behavior of a given system. Consequently, the concept of\u0000typicality, and the associated Asymptotic Equipartition Property allow for a\u0000drastic reduction of the degrees of freedom needed for system's statistical\u0000description. The mathematical rationale for such a simplification in the\u0000description is due to the phenomenon of concentration of measure. The later\u0000emerges for equilibrium configurations thanks to very strict constraints on the\u0000underlying dynamics, such as weekly interacting and (almost) independent system\u0000constituents. The question naturally arises as to whether the concentration of\u0000measure and related typicality considerations can be extended and applied to\u0000more general complex systems, and if so, what mathematical structure can be\u0000expected in the ensuing generalized thermodynamics. In this paper we illustrate\u0000the relevance of the concept of typicality in the toy model context of the\u0000\"thermalized\" coin and show how this leads naturally to Shannon entropy. We\u0000also show an intriguing connection: The characterization of typical sets in\u0000terms of Renyi and Tsallis entropies naturally leads to the free energy and\u0000partition function, respectively, and makes their relationship explicit.\u0000Finally, we propose potential ways to generalize the concept of typicality to\u0000systems where the standard microscopic assumptions do not hold.","PeriodicalId":501082,"journal":{"name":"arXiv - MATH - Information Theory","volume":"35 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-09-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142217198","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}