{"title":"Non-asymptotic model selection for models of network data with parameter vectors of increasing dimension","authors":"Sean Eli , Michael Schweinberger","doi":"10.1016/j.jspi.2024.106173","DOIUrl":"https://doi.org/10.1016/j.jspi.2024.106173","url":null,"abstract":"<div><p>Model selection for network data is an open area of research. Using the <span><math><mi>β</mi></math></span>-model as a convenient starting point, we propose a simple and non-asymptotic approach to model selection of <span><math><mi>β</mi></math></span>-models with and without constraints. Simulations indicate that the proposed model selection approach selects the data-generating model with high probability, in contrast to classical and extended Bayesian Information Criteria. We conclude with an application to the Enron email network, which has 181,831 connections among 36,692 employees.</p></div>","PeriodicalId":50039,"journal":{"name":"Journal of Statistical Planning and Inference","volume":"233 ","pages":"Article 106173"},"PeriodicalIF":0.9,"publicationDate":"2024-04-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140536570","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Hermite regression estimation in noisy convolution model","authors":"Ousmane Sacko","doi":"10.1016/j.jspi.2024.106168","DOIUrl":"https://doi.org/10.1016/j.jspi.2024.106168","url":null,"abstract":"<div><p>In this paper, we consider the following regression model: <span><math><mrow><mi>y</mi><mrow><mo>(</mo><mi>k</mi><mi>T</mi><mo>/</mo><mi>n</mi><mo>)</mo></mrow><mo>=</mo><mi>f</mi><mo>⋆</mo><mi>g</mi><mrow><mo>(</mo><mi>k</mi><mi>T</mi><mo>/</mo><mi>n</mi><mo>)</mo></mrow><mo>+</mo><msub><mrow><mi>ɛ</mi></mrow><mrow><mi>k</mi></mrow></msub><mo>,</mo><mi>k</mi><mo>=</mo><mo>−</mo><mi>n</mi><mo>,</mo><mo>…</mo><mo>,</mo><mi>n</mi><mo>−</mo><mn>1</mn></mrow></math></span>, <span><math><mi>T</mi></math></span> fixed, where <span><math><mi>g</mi></math></span> is known and <span><math><mi>f</mi></math></span> is the unknown function to be estimated. The errors <span><math><msub><mrow><mrow><mo>(</mo><msub><mrow><mi>ɛ</mi></mrow><mrow><mi>k</mi></mrow></msub><mo>)</mo></mrow></mrow><mrow><mo>−</mo><mi>n</mi><mo>≤</mo><mi>k</mi><mo>≤</mo><mi>n</mi><mo>−</mo><mn>1</mn></mrow></msub></math></span> are independent and identically distributed centered with finite known variance. Two adaptive estimation methods for <span><math><mi>f</mi></math></span> are considered by exploiting the properties of the Hermite basis. We study the quadratic risk of each estimator. If <span><math><mi>f</mi></math></span> belongs to Sobolev regularity spaces, we derive rates of convergence. Adaptive procedures to select the relevant parameter inspired by the Goldenshluger and Lepski method are proposed and we prove that the resulting estimators satisfy oracle inequalities for sub-Gaussian <span><math><mi>ɛ</mi></math></span>’s. Finally, we illustrate numerically these approaches.</p></div>","PeriodicalId":50039,"journal":{"name":"Journal of Statistical Planning and Inference","volume":"233 ","pages":"Article 106168"},"PeriodicalIF":0.9,"publicationDate":"2024-03-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140350068","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"How many neurons do we need? A refined analysis for shallow networks trained with gradient descent","authors":"Mike Nguyen, Nicole Mücke","doi":"10.1016/j.jspi.2024.106169","DOIUrl":"https://doi.org/10.1016/j.jspi.2024.106169","url":null,"abstract":"<div><p>We analyze the generalization properties of two-layer neural networks in the neural tangent kernel (NTK) regime, trained with gradient descent (GD). For early stopped GD we derive fast rates of convergence that are known to be minimax optimal in the framework of non-parametric regression in reproducing kernel Hilbert spaces. On our way, we precisely keep track of the number of hidden neurons required for generalization and improve over existing results. We further show that the weights during training remain in a vicinity around initialization, the radius being dependent on structural assumptions such as degree of smoothness of the regression function and eigenvalue decay of the integral operator associated to the NTK.</p></div>","PeriodicalId":50039,"journal":{"name":"Journal of Statistical Planning and Inference","volume":"233 ","pages":"Article 106169"},"PeriodicalIF":0.9,"publicationDate":"2024-03-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0378375824000260/pdfft?md5=7d38fc782951295689c7e96160824723&pid=1-s2.0-S0378375824000260-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140308610","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A class of mixed-level uniform designs generated by code mapping","authors":"Liuping Hu , Zujun Ou , Hong Qin","doi":"10.1016/j.jspi.2024.106166","DOIUrl":"https://doi.org/10.1016/j.jspi.2024.106166","url":null,"abstract":"<div><p>Literature reviews reveal that there is a very close connection between experimental design and coding theory. Based on a code mapping transformation, this paper provides a new method to construct a class of mixed designs with two- and four-level. A general construction method is described and some theoretical results of obtained designs are given. Analytic connections are established between the generated and the initial designs in terms of aberration criteria and discrepancies. Sharp lower bounds of the wrap-around <span><math><msub><mrow><mi>L</mi></mrow><mrow><mn>2</mn></mrow></msub></math></span>- and Lee discrepancies are obtained and used as the benchmarks to measure the uniformity of the generated designs. Examples are provided to illustrate the effectiveness of the construction and lend our results further support.</p></div>","PeriodicalId":50039,"journal":{"name":"Journal of Statistical Planning and Inference","volume":"233 ","pages":"Article 106166"},"PeriodicalIF":0.9,"publicationDate":"2024-03-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140341497","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Robust estimation of a regression function in exponential families","authors":"Yannick Baraud, Juntong Chen","doi":"10.1016/j.jspi.2024.106167","DOIUrl":"https://doi.org/10.1016/j.jspi.2024.106167","url":null,"abstract":"<div><p>We observe <span><math><mi>n</mi></math></span> pairs of independent (but not necessarily i.i.d.) random variables <span><math><mrow><msub><mrow><mi>X</mi></mrow><mrow><mn>1</mn></mrow></msub><mo>=</mo><mrow><mo>(</mo><msub><mrow><mi>W</mi></mrow><mrow><mn>1</mn></mrow></msub><mo>,</mo><msub><mrow><mi>Y</mi></mrow><mrow><mn>1</mn></mrow></msub><mo>)</mo></mrow><mo>,</mo><mo>…</mo><mo>,</mo><msub><mrow><mi>X</mi></mrow><mrow><mi>n</mi></mrow></msub><mo>=</mo><mrow><mo>(</mo><msub><mrow><mi>W</mi></mrow><mrow><mi>n</mi></mrow></msub><mo>,</mo><msub><mrow><mi>Y</mi></mrow><mrow><mi>n</mi></mrow></msub><mo>)</mo></mrow></mrow></math></span> and tackle the problem of estimating the conditional distributions <span><math><mrow><msubsup><mrow><mi>Q</mi></mrow><mrow><mi>i</mi></mrow><mrow><mo>⋆</mo></mrow></msubsup><mrow><mo>(</mo><msub><mrow><mi>w</mi></mrow><mrow><mi>i</mi></mrow></msub><mo>)</mo></mrow></mrow></math></span> of <span><math><msub><mrow><mi>Y</mi></mrow><mrow><mi>i</mi></mrow></msub></math></span> given <span><math><mrow><msub><mrow><mi>W</mi></mrow><mrow><mi>i</mi></mrow></msub><mo>=</mo><msub><mrow><mi>w</mi></mrow><mrow><mi>i</mi></mrow></msub></mrow></math></span> for all <span><math><mrow><mi>i</mi><mo>∈</mo><mrow><mo>{</mo><mn>1</mn><mo>,</mo><mo>…</mo><mo>,</mo><mi>n</mi><mo>}</mo></mrow></mrow></math></span>. Even though these might not be true, we base our estimator on the assumptions that the data are i.i.d. and the conditional distributions of <span><math><msub><mrow><mi>Y</mi></mrow><mrow><mi>i</mi></mrow></msub></math></span> given <span><math><mrow><msub><mrow><mi>W</mi></mrow><mrow><mi>i</mi></mrow></msub><mo>=</mo><msub><mrow><mi>w</mi></mrow><mrow><mi>i</mi></mrow></msub></mrow></math></span> belong to a one parameter exponential family <span><math><mover><mrow><mi>Q</mi></mrow><mo>¯</mo></mover></math></span> with parameter space given by an interval <span><math><mi>I</mi></math></span>. More precisely, we pretend that these conditional distributions take the form <span><math><mrow><msub><mrow><mi>Q</mi></mrow><mrow><mi>θ</mi><mrow><mo>(</mo><msub><mrow><mi>w</mi></mrow><mrow><mi>i</mi></mrow></msub><mo>)</mo></mrow></mrow></msub><mo>∈</mo><mover><mrow><mi>Q</mi></mrow><mo>¯</mo></mover></mrow></math></span> for some <span><math><mi>θ</mi></math></span> that belongs to a VC-class <span><math><mover><mrow><mi>Θ</mi></mrow><mo>¯</mo></mover></math></span> of functions with values in <span><math><mi>I</mi></math></span>. For each <span><math><mrow><mi>i</mi><mo>∈</mo><mrow><mo>{</mo><mn>1</mn><mo>,</mo><mo>…</mo><mo>,</mo><mi>n</mi><mo>}</mo></mrow></mrow></math></span>, we estimate <span><math><mrow><msubsup><mrow><mi>Q</mi></mrow><mrow><mi>i</mi></mrow><mrow><mo>⋆</mo></mrow></msubsup><mrow><mo>(</mo><msub><mrow><mi>w</mi></mrow><mrow><mi>i</mi></mrow></msub><mo>)</mo></mrow></mrow></math></span> by a distribution of the same form, i.e. <span><math><mrow><msub><mrow><mi>Q</mi></mrow><mrow><mover><mrow><mi>θ</mi></","PeriodicalId":50039,"journal":{"name":"Journal of Statistical Planning and Inference","volume":"233 ","pages":"Article 106167"},"PeriodicalIF":0.9,"publicationDate":"2024-03-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140308611","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Measures of conditional dependence for nonlinearity, asymmetry and beyond","authors":"Lianyan Fu , Luyang Zhang","doi":"10.1016/j.jspi.2024.106165","DOIUrl":"https://doi.org/10.1016/j.jspi.2024.106165","url":null,"abstract":"<div><p>Detecting the correlation between two random variables is widely used in many empirical problems in economics. Among them, Pearson’s correlation can be used to quantify the degree of dependence between variables. However, it cannot handle asymmetric correlations. To deal with this situation, we proposed a pair of widely applicable measures of conditional dependence (<span><math><mrow><mi>M</mi><mi>C</mi><mi>D</mi><mi>s</mi></mrow></math></span>), which can not only account for the asymmetry but also the linear or nonlinear conditional dependencies in the presence of multiple variables. We give instances: when the paired measures are the same, resulting in symmetric correlation measures that are equivalent to the square of the Pearson coefficient; when no condition variables are given, <span><math><mrow><mi>M</mi><mi>C</mi><mi>D</mi><mi>s</mi></mrow></math></span> are used to assess the relationship between two variables. Consequently, Pearson’s correlation is a particular instance of <span><math><mrow><mi>M</mi><mi>C</mi><mi>D</mi><mi>s</mi></mrow></math></span>. Theoretical attributes of <span><math><mrow><mi>M</mi><mi>C</mi><mi>D</mi><mi>s</mi></mrow></math></span> show that they have wide applicability. In statistical inference, we develop the joint asymptotics of kernel-based estimators for <span><math><mrow><mi>M</mi><mi>C</mi><mi>D</mi><mi>s</mi></mrow></math></span>, which can be applied to determine whether two randomly generated variables exhibit symmetric conditional dependence in the presence of confounding variables. In the simulation, we verify the efficacy of the proposed <span><math><mrow><mi>M</mi><mi>C</mi><mi>D</mi><mi>s</mi></mrow></math></span>. Then we use real data to analyze the asymmetric impact of <span><math><mrow><mi>M</mi><mi>C</mi><mi>D</mi><mi>s</mi></mrow></math></span> on stock market movements.</p></div>","PeriodicalId":50039,"journal":{"name":"Journal of Statistical Planning and Inference","volume":"232 ","pages":"Article 106165"},"PeriodicalIF":0.9,"publicationDate":"2024-03-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140163834","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Multiple testing in genome-wide association studies via hierarchical hidden Markov models","authors":"Pengfei Wang, Zhaofeng Tian","doi":"10.1016/j.jspi.2024.106161","DOIUrl":"https://doi.org/10.1016/j.jspi.2024.106161","url":null,"abstract":"<div><p>Problems of large-scale multiple testing are often encountered in modern scientific research. Conventional multiple testing procedures usually suffer considerable loss of testing efficiency when correlations among tests are ignored. In fact, appropriate use of correlation information not only enhances the efficacy of the testing procedure, but also improves the interpretability of the results. Since the disease- or trait-related single nucleotide polymorphisms (SNPs) tend to be clustered and exhibit serial correlations, hidden Markov model (HMM) based multiple testing procedures have been successfully applied in genome-wide association studies (GWAS). However, modeling the entire chromosome using a single HMM is somewhat rough. To overcome this issue, this paper employs the hierarchical hidden Markov model (HHMM) to describe local correlations among tests, and develops a multiple testing procedure that can automatically divide different class of chromosome regions, while taking into account local correlations among tests. We first propose an oracle procedure that is shown theoretically to be valid, and in fact optimal in some sense. We then develop a date-driven procedure to mimic the oracle version. Extensive simulations and a real data example show that the novel multiple testing procedure outperforms its competitors.</p></div>","PeriodicalId":50039,"journal":{"name":"Journal of Statistical Planning and Inference","volume":"232 ","pages":"Article 106161"},"PeriodicalIF":0.9,"publicationDate":"2024-02-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140041559","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Wanfeng Liang , Yuhao Zhang , Jiyang Wang , Yue Wu , Xiaoyan Ma
{"title":"A new approach for ultrahigh dimensional precision matrix estimation","authors":"Wanfeng Liang , Yuhao Zhang , Jiyang Wang , Yue Wu , Xiaoyan Ma","doi":"10.1016/j.jspi.2024.106164","DOIUrl":"https://doi.org/10.1016/j.jspi.2024.106164","url":null,"abstract":"<div><p>The modified Cholesky decomposition (MCD) method is commonly used in precision matrix estimation assuming that the random variables have a specified order. In this paper, we develop a permutation-based refitted cross validation (PRCV) estimation procedure for ultrahigh dimensional precision matrix based on the MCD, which does not rely on the order of variables. The consistency of the proposed estimator is established under the Frobenius norm without normal distribution assumption. Simulation studies present satisfactory performance of in various scenarios. The proposed method is also applied to analyze a real data. We provide the complete code at <span>https://github.com/lwfwhunanhero/PRCV</span><svg><path></path></svg>.</p></div>","PeriodicalId":50039,"journal":{"name":"Journal of Statistical Planning and Inference","volume":"232 ","pages":"Article 106164"},"PeriodicalIF":0.9,"publicationDate":"2024-02-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139999247","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Deep learning for ψ-weakly dependent processes","authors":"William Kengne, Modou Wade","doi":"10.1016/j.jspi.2024.106163","DOIUrl":"https://doi.org/10.1016/j.jspi.2024.106163","url":null,"abstract":"<div><p>In this paper, we perform deep neural networks for learning stationary <span><math><mi>ψ</mi></math></span>-weakly dependent processes. Such weak-dependence property includes a class of weak dependence conditions such as mixing, association<span><math><mrow><mo>⋯</mo><mspace></mspace></mrow></math></span> and the setting considered here covers many commonly used situations such as: regression estimation, time series prediction, time series classification<span><math><mrow><mo>⋯</mo><mspace></mspace></mrow></math></span> The consistency of the empirical risk minimization algorithm in the class of deep neural networks predictors is established. We achieve the generalization bound and obtain an asymptotic learning rate, which is less than <span><math><mrow><mi>O</mi><mrow><mo>(</mo><msup><mrow><mi>n</mi></mrow><mrow><mo>−</mo><mn>1</mn><mo>/</mo><mi>α</mi></mrow></msup><mo>)</mo></mrow></mrow></math></span>, for all <span><math><mrow><mi>α</mi><mo>></mo><mn>2</mn></mrow></math></span>. A bound of the excess risk, for a wide class of target functions, is also derived. Applications to binary time series classification and prediction in affine causal models with exogenous covariates are carried out. Some simulation results are provided, as well as an application to the US recession data.</p></div>","PeriodicalId":50039,"journal":{"name":"Journal of Statistical Planning and Inference","volume":"232 ","pages":"Article 106163"},"PeriodicalIF":0.9,"publicationDate":"2024-02-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139999248","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"D4R: Doubly robust reduced rank regression in high dimension","authors":"Xiaoyan Ma , Lili Wei , Wanfeng Liang","doi":"10.1016/j.jspi.2024.106162","DOIUrl":"https://doi.org/10.1016/j.jspi.2024.106162","url":null,"abstract":"<div><p>In this paper, we study high-dimensional reduced rank regression and propose a doubly robust procedure, called <span><math><mi>D4R</mi></math></span>, meaning concurrent robustness to both outliers in predictors and heavy-tailed random noise. The proposed method uses the composite gradient descent based algorithm to solve the nonconvex optimization problem resulting from combining Tukey’s biweight loss with spectral regularization. Both theoretical and numerical properties of <span><math><mi>D4R</mi></math></span> are investigated. We establish non-asymptotic estimation error bounds under both the Frobenius norm and the nuclear norm in the high-dimensional setting. Simulation studies and real example show that the performance of <span><math><mi>D4R</mi></math></span> is better than that of several existing estimation methods.</p></div>","PeriodicalId":50039,"journal":{"name":"Journal of Statistical Planning and Inference","volume":"232 ","pages":"Article 106162"},"PeriodicalIF":0.9,"publicationDate":"2024-02-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139985505","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}