Maren Hackenberg, Michelle Pfaffenlehner, Max Behrens, Astrid Pechmann, Janbernd Kirschner, Harald Binder
{"title":"Investigating a Domain Adaptation Approach for Integrating Different Measurement Instruments in a Longitudinal Clinical Registry","authors":"Maren Hackenberg, Michelle Pfaffenlehner, Max Behrens, Astrid Pechmann, Janbernd Kirschner, Harald Binder","doi":"10.1002/bimj.70023","DOIUrl":"10.1002/bimj.70023","url":null,"abstract":"<p>In a longitudinal clinical registry, different measurement instruments might have been used for assessing individuals at different time points. To combine them, we investigate deep learning techniques for obtaining a joint latent representation, to which the items of different measurement instruments are mapped. This corresponds to domain adaptation, an established concept in computer science for image data. Using the proposed approach as an example, we evaluate the potential of domain adaptation in a longitudinal cohort setting with a rather small number of time points, motivated by an application with different motor function measurement instruments in a registry of spinal muscular atrophy (SMA) patients. There, we model trajectories in the latent representation by ordinary differential equations (ODEs), where person-specific ODE parameters are inferred from baseline characteristics. The goodness of fit and complexity of the ODE solutions then allow to judge the measurement instrument mappings. We subsequently explore how alignment can be improved by incorporating corresponding penalty terms into model fitting. To systematically investigate the effect of differences between measurement instruments, we consider several scenarios based on modified SMA data, including scenarios where a mapping should be feasible in principle and scenarios where no perfect mapping is available. While misalignment increases in more complex scenarios, some structure is still recovered, even if the availability of measurement instruments depends on patient state. A reasonable mapping is feasible also in the more complex real SMA data set. These results indicate that domain adaptation might be more generally useful in statistical modeling for longitudinal registry data.</p>","PeriodicalId":55360,"journal":{"name":"Biometrical Journal","volume":"67 1","pages":""},"PeriodicalIF":1.3,"publicationDate":"2024-12-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/bimj.70023","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142857076","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"生物学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Milena Wünsch, Christina Sauer, Moritz Herrmann, Ludwig Christian Hinske, Anne-Laure Boulesteix
{"title":"To Tweak or Not to Tweak. How Exploiting Flexibilities in Gene Set Analysis Leads to Overoptimism","authors":"Milena Wünsch, Christina Sauer, Moritz Herrmann, Ludwig Christian Hinske, Anne-Laure Boulesteix","doi":"10.1002/bimj.70016","DOIUrl":"10.1002/bimj.70016","url":null,"abstract":"<p>Gene set analysis, a popular approach for analyzing high-throughput gene expression data, aims to identify sets of genes that show enriched expression patterns between two conditions. In addition to the multitude of methods available for this task, users are typically left with many options when creating the required input and specifying the internal parameters of the chosen method. This flexibility can lead to uncertainty about the “right” choice, further reinforced by a lack of evidence-based guidance. Especially when their statistical experience is scarce, this uncertainty might entice users to produce preferable results using a “trial-and-error” approach. While it may seem unproblematic at first glance, this practice can be viewed as a form of “cherry-picking” and cause an optimistic bias, rendering the results nonreplicable on independent data. After this problem has attracted a lot of attention in the context of classical hypothesis testing, we now aim to raise awareness of such overoptimism in the different and more complex context of gene set analyses. We mimic a hypothetical researcher who systematically selects the analysis variants yielding their preferred results, thereby considering three distinct goals they might pursue. Using a selection of popular gene set analysis methods, we tweak the results in this way for two frequently used benchmark gene expression data sets. Our study indicates that the potential for overoptimism is particularly high for a group of methods frequently used despite being commonly criticized. We conclude by providing practical recommendations to counter overoptimism in research findings in gene set analysis and beyond.</p>","PeriodicalId":55360,"journal":{"name":"Biometrical Journal","volume":"67 1","pages":""},"PeriodicalIF":1.3,"publicationDate":"2024-12-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/bimj.70016","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142857080","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"生物学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Dominic Edelmann, Tobias Terzer, Peter Horak, Richard Schlenk, Axel Benner
{"title":"The Progression-Free-Survival Ratio in Molecularly Aided Tumor Trials: A Critical Examination of Current Practice and Suggestions for Alternative Methods","authors":"Dominic Edelmann, Tobias Terzer, Peter Horak, Richard Schlenk, Axel Benner","doi":"10.1002/bimj.70028","DOIUrl":"10.1002/bimj.70028","url":null,"abstract":"<p>The progression-free-survival ratio is a popular endpoint in oncology trials, which is frequently applied to evaluate the efficacy of molecularly targeted treatments in late-stage patients. Using elementary calculations and simulations, numerous shortcomings of the current methodology are pointed out. As a remedy to these shortcomings, an alternative methodology is proposed, using a marginal Cox model or a marginal accelerated failure time model for clustered time-to-event data. Using comprehensive simulations, it is shown that this methodology outperforms existing methods in settings where the intrapatient correlation is low to moderate. The performance of the model is further demonstrated in a real data example from a molecularly aided tumor trial. Sample size considerations are discussed.</p>","PeriodicalId":55360,"journal":{"name":"Biometrical Journal","volume":"67 1","pages":""},"PeriodicalIF":1.3,"publicationDate":"2024-12-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/bimj.70028","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142848551","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"生物学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Jeppe Ekstrand Halkjær Madsen, Thomas Delvin, Thomas Scheike, Christian Pipper
{"title":"A Principled Approach to Adjust for Unmeasured Time-Stable Confounding of Supervised Treatment","authors":"Jeppe Ekstrand Halkjær Madsen, Thomas Delvin, Thomas Scheike, Christian Pipper","doi":"10.1002/bimj.70026","DOIUrl":"10.1002/bimj.70026","url":null,"abstract":"<div>\u0000 \u0000 <p>We propose a novel method to adjust for unmeasured time-stable confounding when the time between consecutive treatment administrations is fixed. We achieve this by focusing on a new-user cohort. Furthermore, we envisage that all time-stable confounding goes through the potential time on treatment as dictated by the disease condition at the initiation of treatment. Following this logic, we may eliminate all unmeasured time-stable confounding by adjusting for the potential time on treatment. A challenge with this approach is that right censoring of the potential time on treatment occurs when treatment is terminated at the time of the event of interest, for example, if the event of interest is death. We show how this challenge may be solved by means of the expectation-maximization algorithm without imposing any further assumptions on the distribution of the potential time on treatment. The usefulness of the methodology is illustrated in a simulation study. We also apply the methodology to investigate the effect of depression/anxiety drugs on subsequent poisoning by other medications in the Danish population by means of national registries. We find a protective effect of treatment with selective serotonin reuptake inhibitors on the risk of poisoning by various medications (1- year risk difference of approximately <span></span><math>\u0000 <semantics>\u0000 <mrow>\u0000 <mo>−</mo>\u0000 <mn>3</mn>\u0000 <mo>%</mo>\u0000 </mrow>\u0000 <annotation>$-3%$</annotation>\u0000 </semantics></math>) and a standard Cox model analysis shows a harming effect (1-year risk difference of approximately <span></span><math>\u0000 <semantics>\u0000 <mrow>\u0000 <mn>2</mn>\u0000 <mo>%</mo>\u0000 </mrow>\u0000 <annotation>$2%$</annotation>\u0000 </semantics></math>), which is consistent with what we would expect due to confounding by indication. Unmeasured time-stable confounding can be entirely adjusted for when the time between consecutive treatment administrations is fixed.</p></div>","PeriodicalId":55360,"journal":{"name":"Biometrical Journal","volume":"67 1","pages":""},"PeriodicalIF":1.3,"publicationDate":"2024-12-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142840159","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"生物学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Assessing Balance of Baseline Time-Dependent Covariates via the Fréchet Distance","authors":"Mireya Díaz","doi":"10.1002/bimj.70024","DOIUrl":"10.1002/bimj.70024","url":null,"abstract":"<div>\u0000 \u0000 <p>Assessment of covariate balance is a key step when performing comparisons between groups particularly in real-world data. We generally evaluate it on baseline covariates, but rarely on longitudinal ones prior to a management decision. We could use pointwise standardized mean differences, standardized differences of slopes, or weights from the model for such purpose. Pointwise differences could be cumbersome for densely sampled longitudinal markers and/or measured at different points. Slopes are suitable for linear or transformable models but not for more complex curves. Weights do not identify the specific covariate(s) responsible for imbalances. This work presents the Fréchet distance as a viable alternative to assess balance of time-dependent covariates. A set of linear and nonlinear curves for which their standardized difference or differences in functional parameters were within 10% sought to identify the Fréchet distance equivalent to this threshold. This threshold is dependent on the level of noise present and thus within group heterogeneity and error variance are needed for its interpretation. Applied to a set of real curves representing the monthly trajectory of hemoglobin A1c from diabetic patients showed that the curves in the two groups were not balanced at the 10% mark. A Beta distribution represents the Fréchet distance distribution reasonably well in most scenarios. This assessment of covariate balance provides the following advantages: It can handle curves of different lengths, shapes, and arbitrary time points. Future work includes examining the utility of this measure under within-series missingness, within-group heterogeneity, its comparison with other approaches, and asymptotics.</p>\u0000 </div>","PeriodicalId":55360,"journal":{"name":"Biometrical Journal","volume":"67 1","pages":""},"PeriodicalIF":1.3,"publicationDate":"2024-12-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142840161","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"生物学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Alexandra Erdmann, Jan Beyersmann, Kaspar Rufibach
{"title":"Oncology Clinical Trial Design Planning Based on a Multistate Model That Jointly Models Progression-Free and Overall Survival Endpoints","authors":"Alexandra Erdmann, Jan Beyersmann, Kaspar Rufibach","doi":"10.1002/bimj.70017","DOIUrl":"10.1002/bimj.70017","url":null,"abstract":"<p>When planning an oncology clinical trial, the usual approach is to assume proportional hazards and even an exponential distribution for time-to-event endpoints. Often, besides the gold-standard endpoint overall survival (OS), progression-free survival (PFS) is considered as a second confirmatory endpoint. We use a survival multistate model to jointly model these two endpoints and find that neither exponential distribution nor proportional hazards will typically hold for both endpoints simultaneously. The multistate model provides a stochastic process approach to model the dependency of such endpoints neither requiring latent failure times nor explicit dependency modeling such as copulae. We use the multistate model framework to simulate clinical trials with endpoints OS and PFS and show how design planning questions can be answered using this approach. In particular, nonproportional hazards for at least one of the endpoints are a consequence of OS and PFS being dependent and are naturally modeled to improve planning. We then illustrate how clinical trial design can be based on simulations from a multistate model. Key applications are coprimary endpoints and group-sequential designs. Simulations for these applications show that the standard simplifying approach may very well lead to underpowered or overpowered clinical trials. Our approach is quite general and can be extended to more complex trial designs, further endpoints, and other therapeutic areas. An R package is available on CRAN.</p>","PeriodicalId":55360,"journal":{"name":"Biometrical Journal","volume":"67 1","pages":""},"PeriodicalIF":1.3,"publicationDate":"2024-12-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/bimj.70017","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142840153","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"生物学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Test Statistics and Statistical Inference for Data With Informative Cluster Sizes","authors":"Soyoung Kim, Michael J. Martens, Kwang Woo Ahn","doi":"10.1002/bimj.70021","DOIUrl":"10.1002/bimj.70021","url":null,"abstract":"<div>\u0000 \u0000 <p>In biomedical studies, investigators often encounter clustered data. The cluster sizes are said to be informative if the outcome depends on the cluster size. Ignoring informative cluster sizes in the analysis leads to biased parameter estimation in marginal and mixed-effect regression models. Several methods to analyze data with informative cluster sizes have been proposed; however, methods to test the informativeness of the cluster sizes are limited, particularly for the marginal model. In this paper, we propose a score test and a Wald test to examine the informativeness of the cluster sizes for a generalized linear model, a Cox model, and a proportional subdistribution hazards model. Statistical inference can be conducted through weighted estimating equations. The simulation results show that both tests control Type I error rates well, but the score test has higher power than the Wald test for right-censored data while the power of the Wald test is generally higher than the score test for the binary outcome. We apply the Wald and score tests to hematopoietic cell transplant data and compare regression analysis results with/without adjusting for informative cluster sizes.</p></div>","PeriodicalId":55360,"journal":{"name":"Biometrical Journal","volume":"67 1","pages":""},"PeriodicalIF":1.3,"publicationDate":"2024-12-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142840154","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"生物学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Best Subset Solution Path for Linear Dimension Reduction Models Using Continuous Optimization","authors":"Benoit Liquet, Sarat Moka, Samuel Muller","doi":"10.1002/bimj.70015","DOIUrl":"10.1002/bimj.70015","url":null,"abstract":"<div>\u0000 \u0000 <p>The selection of best variables is a challenging problem in supervised and unsupervised learning, especially in high-dimensional contexts where the number of variables is usually much larger than the number of observations. In this paper, we focus on two multivariate statistical methods: principal components analysis and partial least squares. Both approaches are popular linear dimension-reduction methods with numerous applications in several fields including in genomics, biology, environmental science, and engineering. In particular, these approaches build principal components, new variables that are combinations of all the original variables. A main drawback of principal components is the difficulty to interpret them when the number of variables is large. To define principal components from the most relevant variables, we propose to cast the best subset solution path method into principal component analysis and partial least square frameworks. We offer a new alternative by exploiting a continuous optimization algorithm for best subset solution path. Empirical studies show the efficacy of our approach for providing the best subset solution path. The usage of our algorithm is further exposed through the analysis of two real data sets. The first data set is analyzed using the principle component analysis while the analysis of the second data set is based on partial least square framework.</p></div>","PeriodicalId":55360,"journal":{"name":"Biometrical Journal","volume":"67 1","pages":""},"PeriodicalIF":1.3,"publicationDate":"2024-12-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142840149","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"生物学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Goodness-of-Fit Testing for a Regression Model With a Doubly Truncated Response","authors":"Jacobo de Uña-Álvarez","doi":"10.1002/bimj.70022","DOIUrl":"10.1002/bimj.70022","url":null,"abstract":"<p>In survival analysis and epidemiology, among other fields, interval sampling is often employed. With interval sampling, the individuals undergoing the event of interest within a calendar time interval are recruited. This results in doubly truncated event times. Double truncation, which may appear with other sampling designs too, induces a selection bias, so ordinary statistical methods are generally inconsistent. In this paper, we introduce goodness-of-fit procedures for a regression model when the response variable is doubly truncated. With this purpose, a marked empirical process based on weighted residuals is constructed and its weak convergence is established. Kolmogorov–Smirnov– and Cramér–von Mises–type tests are consequently derived from such core process, and a bootstrap approximation for their practical implementation is given. The performance of the proposed tests is investigated through simulations. An application to model selection for AIDS incubation time as depending on age at infection is provided.</p>","PeriodicalId":55360,"journal":{"name":"Biometrical Journal","volume":"67 1","pages":""},"PeriodicalIF":1.3,"publicationDate":"2024-12-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/bimj.70022","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142840151","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"生物学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Yujie Zhao, Qi Liu, Linda Z. Sun, Keaven M. Anderson
{"title":"Adjusted Inference for Multiple Testing Procedure in Group-Sequential Designs","authors":"Yujie Zhao, Qi Liu, Linda Z. Sun, Keaven M. Anderson","doi":"10.1002/bimj.70020","DOIUrl":"10.1002/bimj.70020","url":null,"abstract":"<div>\u0000 \u0000 <p>Adjustment of statistical significance levels for repeated analysis in group-sequential trials has been understood for some time. Adjustment accounting for testing multiple hypotheses is also well understood. There is limited research on simultaneously adjusting for both multiple hypothesis testing and repeated analyses of one or more hypotheses. We address this gap by proposing <i>adjusted-sequential p-values</i> that reject when they are less than or equal to the family-wise Type I error rate (FWER). We also propose sequential <span></span><math>\u0000 <semantics>\u0000 <mi>p</mi>\u0000 <annotation>$p$</annotation>\u0000 </semantics></math>-values for intersection hypotheses to compute adjusted-sequential <span></span><math>\u0000 <semantics>\u0000 <mi>p</mi>\u0000 <annotation>$p$</annotation>\u0000 </semantics></math>-values for elementary hypotheses. We demonstrate the application using weighted Bonferroni tests and weighted parametric tests for inference on each elementary hypothesis tested.</p></div>","PeriodicalId":55360,"journal":{"name":"Biometrical Journal","volume":"67 1","pages":""},"PeriodicalIF":1.3,"publicationDate":"2024-12-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142840160","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"生物学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}