Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 20 de 63
Filtrar
1.
Stat Med ; 43(18): 3417-3431, 2024 Aug 15.
Artículo en Inglés | MEDLINE | ID: mdl-38852994

RESUMEN

We investigate the familywise error rate (FWER) for time-to-event endpoints evaluated using a group sequential design with a hierarchical testing procedure for secondary endpoints. We show that, in this setup, the correlation between the log-rank test statistics at interim and at end of study is not congruent with the canonical correlation derived for normal-distributed endpoints. We show, both theoretically and by simulation, that the correlation also depends on the level of censoring, the hazard rates of the endpoints, and the hazard ratio. To optimize operating characteristics in this complex scenario, we propose a simulation-based method to assess the FWER which, better than the alpha-spending approach, can inform the choice of critical values for testing secondary endpoints.


Asunto(s)
Simulación por Computador , Determinación de Punto Final , Humanos , Determinación de Punto Final/métodos , Proyectos de Investigación , Modelos Estadísticos , Modelos de Riesgos Proporcionales , Interpretación Estadística de Datos
2.
Biom J ; 66(3): e2300237, 2024 Apr.
Artículo en Inglés | MEDLINE | ID: mdl-38637319

RESUMEN

In this paper, we consider online multiple testing with familywise error rate (FWER) control, where the probability of committing at least one type I error will remain under control while testing a possibly infinite sequence of hypotheses over time. Currently, adaptive-discard (ADDIS) procedures seem to be the most promising online procedures with FWER control in terms of power. Now, our main contribution is a uniform improvement of the ADDIS principle and thus of all ADDIS procedures. This means, the methods we propose reject as least as much hypotheses as ADDIS procedures and in some cases even more, while maintaining FWER control. In addition, we show that there is no other FWER controlling procedure that enlarges the event of rejecting any hypothesis. Finally, we apply the new principle to derive uniform improvements of the ADDIS-Spending and ADDIS-Graph.


Asunto(s)
Modelos Estadísticos , Probabilidad
3.
Biom J ; 65(8): e2200300, 2023 12.
Artículo en Inglés | MEDLINE | ID: mdl-37789586

RESUMEN

We give a simulation-based method for computing the multiplicity adjusted p-values and critical constants for the Dunnett procedure for comparing treatments with a control under heteroskedasticity. The Welch-Satterthwaite test statistics used in this procedure do not have a simple multivariate t-distribution because their denominators are mixtures of chi-squares and are correlated because of the common control treatment sample variance present in all denominators. The joint distribution of the denominators of the test statistics is approximated by correlated chi-square variables and is generated using a novel algorithm proposed in this paper. This approximation is used to derive critical constants or adjusted p-values. The familywise error rate (FWER) of the proposed method is compared with some existing methods via simulation under different heteroskedastic scenarios. The results show that our proposed method controls the FWER most accurately, whereas other methods are either too conservative or liberal or control the FWER less accurately. The different methods considered are illustrated on a real data set.


Asunto(s)
Algoritmos , Modelos Estadísticos , Simulación por Computador
4.
Biometrics ; 79(4): 2781-2793, 2023 12.
Artículo en Inglés | MEDLINE | ID: mdl-37533251

RESUMEN

We consider the problem of testing multiple null hypotheses, where a decision to reject or retain must be made for each one and embedding incorrect decisions into a real-life context may inflict different losses. We argue that traditional methods controlling the Type I error rate may be too restrictive in this situation and that the standard familywise error rate may not be appropriate. Using a decision-theoretic approach, we define suitable loss functions for a given decision rule, where incorrect decisions can be treated unequally by assigning different loss values. Taking expectation with respect to the sampling distribution of the data allows us to control the familywise expected loss instead of the conventional familywise error rate. Different loss functions can be adopted, and we search for decision rules that satisfy certain optimality criteria within a broad class of decision rules for which the expected loss is bounded by a fixed threshold under any parameter configuration. We illustrate the methods with the problem of establishing efficacy of a new medicinal treatment in non-overlapping subgroups of patients.


Asunto(s)
Proyectos de Investigación , Humanos , Interpretación Estadística de Datos
5.
Neuroinformatics ; 21(3): 601-614, 2023 07.
Artículo en Inglés | MEDLINE | ID: mdl-37314682

RESUMEN

Neurodegeneration measured through volumetry in MRI is recognized as a potential Alzheimer's Disease (AD) biomarker, but its utility is limited by lack of specificity. Quantifying spatial patterns of neurodegeneration on a whole brain scale rather than locally may help address this. In this work, we turn to network based analyses and extend a graph embedding algorithm to study morphometric connectivity from volume-change correlations measured with structural MRI on the timescale of years. We model our data with the multiple random eigengraphs framework, as well as modify and implement a multigraph embedding algorithm proposed earlier to estimate a low dimensional embedding of the networks. Our version of the algorithm guarantees meaningful finite-sample results and estimates maximum likelihood edge probabilities from population-specific network modes and subject-specific loadings. Furthermore, we propose and implement a novel statistical testing procedure to analyze group differences after accounting for confounders and locate significant structures during AD neurodegeneration. Family-wise error rate is controlled at 5% using permutation testing on the maximum statistic. We show that results from our analysis reveal networks dominated by known structures associated to AD neurodegeneration, indicating the framework has promise for studying AD. Furthermore, we find network-structure tuples that are not found with traditional methods in the field.


Asunto(s)
Enfermedad de Alzheimer , Humanos , Enfermedad de Alzheimer/diagnóstico por imagen , Imagen por Resonancia Magnética/métodos , Encéfalo/diagnóstico por imagen , Algoritmos , Biomarcadores
6.
Biometrics ; 79(2): 1103-1113, 2023 06.
Artículo en Inglés | MEDLINE | ID: mdl-35567306

RESUMEN

The Globaltest is a powerful test for the global null hypothesis that there is no association between a group of features and a response of interest, which is popular in pathway testing in metabolomics. Evaluating multiple feature sets, however, requires multiple testing correction. In this paper, we propose a multiple testing method, based on closed testing, specifically designed for the Globaltest. The proposed method controls the familywise error rate simultaneously over all possible feature sets, and therefore allows post hoc inference, that is, the researcher may choose feature sets of interest after seeing the data without jeopardizing error control. To circumvent the exponential computation time of closed testing, we derive a novel shortcut that allows exact closed testing to be performed on the scale of metabolomics data. An R package ctgt is available on comprehensive R archive network for the implementation of the shortcut procedure, with applications on several real metabolomics data examples.


Asunto(s)
Metabolómica
7.
Biometrics ; 79(2): 1114-1118, 2023 06.
Artículo en Inglés | MEDLINE | ID: mdl-35355244

RESUMEN

Hung et al. (2007) considered the problem of controlling the type I error rate for a primary and secondary endpoint in a clinical trial using a gatekeeping approach in which the secondary endpoint is tested only if the primary endpoint crosses its monitoring boundary. They considered a two-look trial and showed by simulation that the naive method of testing the secondary endpoint at full level α at the time the primary endpoint reaches statistical significance does not control the familywise error rate at level α. Tamhane et al. (2010) derived analytic expressions for familywise error rate and power and confirmed the inflated error rate of the naive approach. Nonetheless, many people mistakenly believe that the closure principle can be used to prove that the naive procedure controls the familywise error rate. The purpose of this note is to explain in greater detail why there is a problem with the naive approach and show that the degree of alpha inflation can be as high as that of unadjusted monitoring of a single endpoint.


Asunto(s)
Modelos Estadísticos , Proyectos de Investigación , Humanos , Determinación de Punto Final/métodos , Simulación por Computador , Tamaño de la Muestra
8.
Biom J ; 64(6): 1090-1108, 2022 08.
Artículo en Inglés | MEDLINE | ID: mdl-35426161

RESUMEN

Mediation analysis in high-dimensional settings often involves identifying potential mediators among a large number of measured variables. For this purpose, a two-step familywise error rate procedure called ScreenMin has been recently proposed. In ScreenMin, variables are first screened and only those that pass the screening are tested. The proposed data-independent threshold for selection has been shown to guarantee asymptotic familywise error rate. In this work, we investigate the impact of the threshold on the finite-sample familywise error rate. We derive a power maximizing threshold and show that it is well approximated by an adaptive threshold of Wang et al. (2016, arXiv preprint arXiv:1610.03330). We illustrate the investigated procedures on a case-control study examining the effect of fish intake on the risk of colorectal adenoma. We also apply our procedure in the context of replicability analysis to identify single nucleotide polymorphisms (SNP) associated with crop yield in two distinct environments.


Asunto(s)
Modelos Estadísticos , Animales , Estudios de Casos y Controles
9.
Biometrika ; 109(2): 457-471, 2022 Jun.
Artículo en Inglés | MEDLINE | ID: mdl-38694183

RESUMEN

We consider the problem of multiple hypothesis testing when there is a logical nested structure to the hypotheses. When one hypothesis is nested inside another, the outer hypothesis must be false if the inner hypothesis is false. We model the nested structure as a directed acyclic graph, including chain and tree graphs as special cases. Each node in the graph is a hypothesis and rejecting a node requires also rejecting all of its ancestors. We propose a general framework for adjusting node-level test statistics using the known logical constraints. Within this framework, we study a smoothing procedure that combines each node with all of its descendants to form a more powerful statistic. We prove a broad class of smoothing strategies can be used with existing selection procedures to control the familywise error rate, false discovery exceedance rate, or false discovery rate, so long as the original test statistics are independent under the null. When the null statistics are not independent but are derived from positively-correlated normal observations, we prove control for all three error rates when the smoothing method is arithmetic averaging of the observations. Simulations and an application to a real biology dataset demonstrate that smoothing leads to substantial power gains.

10.
Biometrika ; 108(4): 915-931, 2021 Dec.
Artículo en Inglés | MEDLINE | ID: mdl-34803516

RESUMEN

The familywise error rate has been widely used in genome-wide association studies. With the increasing availability of functional genomics data, it is possible to increase detection power by leveraging these genomic functional annotations. Previous efforts to accommodate covariates in multiple testing focused on false discovery rate control, while covariate-adaptive procedures controlling the familywise error rate remain underdeveloped. Here, we propose a novel covariate-adaptive procedure to control the familywise error rate that incorporates external covariates which are potentially informative of either the statistical power or the prior null probability. An efficient algorithm is developed to implement the proposed method. We prove its asymptotic validity and obtain the rate of convergence through a perturbation-type argument. Our numerical studies show that the new procedure is more powerful than competing methods and maintains robustness across different settings. We apply the proposed approach to the UK Biobank data and analyse 27 traits with 9 million single-nucleotide polymorphisms tested for associations. Seventy-five genomic annotations are used as covariates. Our approach detects more genome-wide significant loci than other methods in 21 out of the 27 traits.

11.
Stat Med ; 40(24): 5333-5350, 2021 10 30.
Artículo en Inglés | MEDLINE | ID: mdl-34636081

RESUMEN

The problem of testing multiple hypotheses using a group sequential procedure often arises in clinical trials. We review several group sequential Holm (GSHM) type procedures proposed in the literature and clarify the relationships between them. In particular, we show which procedures are equivalent or, if different, which are more powerful and what are their pros and cons. We propose a step-up group sequential Hochberg (GSHC) procedure as a reverse application of a particular step-down GSHM procedure. We conducted an extensive simulation study to evaluate the familywise error rate (FWER) and power properties of that GSHM procedure and the GSHC procedure and found that the GSHC procedure controls FWER more closely and is more powerful. All procedures are illustrated with a common numerical example, the data for which are chosen to bring out the differences between them. A real case study is also presented to illustrate application of these procedures. R programs for applying the proposed procedures, additional simulation results, and the proof of the FWER control of the GSHC procedure in a special case are provided in Supplementary Material.


Asunto(s)
Ensayos Clínicos como Asunto , Simulación por Computador
12.
Contemp Clin Trials ; 109: 106538, 2021 10.
Artículo en Inglés | MEDLINE | ID: mdl-34384890

RESUMEN

Master protocols, in particular umbrella trials and platform trials, when evaluating multiple experimental treatments with a common control, could save patient resource, increase trial efficiency, and reduce drug development cost. Compared to the phase 3 platform trials that allow unlimited number of experimental arms to be added, it is more practical for individual companies to evaluate two experimental arms with a common control in an umbrella trial and allow the second experimental arm to be added at a later time. There have been limited research done in this type of trials in terms of statistical properties and guidance. In this article, we present statistical considerations of a phase 3 three-arm umbrella design including Type I error control and power, as well as the optimal allocation ratio. We intend to not only complement the existing literature, but more importantly to provide practical guidance to pave the way for its implementation by individual companies.


Asunto(s)
Ensayos Clínicos Fase III como Asunto , Proyectos de Investigación , Protocolos Clínicos , Desarrollo de Medicamentos , Humanos
13.
J Biopharm Stat ; 31(4): 391-402, 2021 07 04.
Artículo en Inglés | MEDLINE | ID: mdl-33909544

RESUMEN

We introduce an improved Bonferroni method for testing two primary endpoints in clinical trial settings using a new data-adaptive critical value that explicitly incorporates the sample correlation coefficient. Our methodology is developed for the usual Student's t-test statistics for testing the means under normal distributional setting with unknown population correlation and variances. Specifically, we construct a confidence interval for the unknown population correlation and show that the estimated type-1 error rate of the Bonferroni method with the population correlation being estimated by its lower confidence limit can be bounded from above less conservatively than using the traditional Bonferroni upper bound. We also compare the new procedure with other procedures commonly used for the multiple testing problem addressed in this paper.


Asunto(s)
Modelos Estadísticos , Proyectos de Investigación , Interpretación Estadística de Datos , Humanos
14.
Metabolites ; 11(1)2021 Jan 14.
Artículo en Inglés | MEDLINE | ID: mdl-33466792

RESUMEN

Due to the advance in technology, the type of data is getting more complicated and large-scale. To analyze such complex data, more advanced technique is required. In case of omics data from two different groups, it is interesting to find significant biomarkers between two groups while controlling error rate such as false discovery rate (FDR). Over the last few decades, a lot of methods that control local false discovery rate have been developed, ranging from one-dimensional to k-dimensional FDR procedure. For comparison study, we select three of them, which have unique and significant properties: Efron et al. (2001), Ploner et al. (2006), and Kim et al. (2018) in chronological order. The first approach is one-dimensional approach while the other two are two-dimensional ones. Furthermore, we consider two more variants of Ploner's approach. We compare the performance of those methods on both simulated and real data.

15.
F1000Res ; 10: 991, 2021.
Artículo en Inglés | MEDLINE | ID: mdl-36925625

RESUMEN

Background The CONSORT guidelines for clinical trials recommend use of a single primary outcome, to guard against the raised risk of false positive findings when multiple measures are considered. It is, however, possible to include a suite of multiple outcomes in an intervention study, while controlling the familywise error rate, if the criterion for rejecting the null hypothesis specifies that N or more of the outcomes reach an agreed level of statistical significance, where N depends on the total number of outcome measures included in the study, and the correlation between them. Methods Simulations were run, using a conventional null-hypothesis significance testing approach with alpha set at .05, to explore the case when between 2 and 12 outcome measures are included to compare two groups, with average correlation between measures ranging from zero to .8, and true effect size ranging from 0 to .7. In step 1, a table is created giving the minimum N significant outcomes (MinNSig) that is required for a given set of outcome measures to control the familywise error rate at 5%. In step 2, data are simulated using MinNSig values for each set of correlated outcomes and the resulting proportion of significant results is computed for different sample sizes,correlations, and effect sizes. Results The Adjust NVar approach can achieve a more efficient trade-off between power and type I error rate than use of a single outcome when there are three or more moderately intercorrelated outcome variables. Conclusions Where it is feasible to have a suite of moderately correlated outcome measures, then this might be a more efficient approach than reliance on a single primary outcome measure in an intervention study. In effect, it builds in an internal replication to the study. This approach can also be used to evaluate published intervention studies.

16.
J Biopharm Stat ; 30(6): 1077-1090, 2020 11 01.
Artículo en Inglés | MEDLINE | ID: mdl-32990148

RESUMEN

This paper provides in-depth discussion about different types of error generated in platform trials with a common control arm, and how they compare to the ones arisen from standard independent trials. We provide our views on some of the popular "myths" associated with such design, under the frequentist framework. It is found that platform trial generally performs quite well in terms of type I error rate, false discovery rate, and power. In most cases, these operating characteristics of a platform trial are comparable to or even better than running individual trials.


Asunto(s)
Proyectos de Investigación , Interpretación Estadística de Datos , Humanos
17.
Stat Med ; 39(26): 3772-3786, 2020 11 20.
Artículo en Inglés | MEDLINE | ID: mdl-32706424

RESUMEN

Clinical trials routinely involve multiple hypothesis testing. The closed testing procedure (CTP) is a fundamental principle in testing multiple hypotheses. This article presents an improved CTP in which intersection hypotheses can be tested at a level greater than α such that the control of the familywise error rate at level α remains. Consequently, our method uniformly improves the power of discovering false hypotheses over the original CTP. We illustrate that an improvement by our method exists for many commonly used tests. An empirical study on the effectiveness of a glucose-lowering drug is provided.


Asunto(s)
Ensayos Clínicos como Asunto , Proyectos de Investigación , Humanos
18.
Stat Med ; 39(23): 3135-3155, 2020 10 15.
Artículo en Inglés | MEDLINE | ID: mdl-32557848

RESUMEN

When simultaneously testing multiple hypotheses, the usual approach in the context of confirmatory clinical trials is to control the familywise error rate (FWER), which bounds the probability of making at least one false rejection. In many trial settings, these hypotheses will additionally have a hierarchical structure that reflects the relative importance and links between different clinical objectives. The graphical approach of Bretz et al (2009) is a flexible and easily communicable way of controlling the FWER while respecting complex trial objectives and multiple structured hypotheses. However, the FWER can be a very stringent criterion that leads to procedures with low power, and may not be appropriate in exploratory trial settings. This motivates controlling generalized error rates, particularly when the number of hypotheses tested is no longer small. We consider the generalized familywise error rate (k-FWER), which is the probability of making k or more false rejections, as well as the tail probability of the false discovery proportion (FDP), which is the probability that the proportion of false rejections is greater than some threshold. We also consider asymptotic control of the false discovery rate, which is the expectation of the FDP. In this article, we show how to control these generalized error rates when using the graphical approach and its extensions. We demonstrate the utility of the resulting graphical procedures on three clinical trial case studies.


Asunto(s)
Proyectos de Investigación , Humanos , Probabilidad
19.
Stat Med ; 39(9): 1407-1413, 2020 04 30.
Artículo en Inglés | MEDLINE | ID: mdl-32106332

RESUMEN

Multiple comparison adjustments have a long history, yet confusion remains about which procedures control type 1 error rate in a strong sense and how to show this. Part of the confusion stems from a powerful technique called the closed testing principle, whose statement is deceptively simple, but is sometimes misinterpreted. This primer presents a straightforward way to think about multiplicity adjustment.

20.
BMC Cancer ; 20(1): 80, 2020 Jan 31.
Artículo en Inglés | MEDLINE | ID: mdl-32005187

RESUMEN

BACKGROUND: Multi-arm designs provide an effective means of evaluating several treatments within the same clinical trial. Given the large number of treatments now available for testing in many disease areas, it has been argued that their utilisation should increase. However, for any given clinical trial there are numerous possible multi-arm designs that could be used, and choosing between them can be a difficult task. This task is complicated further by a lack of available easy-to-use software for designing multi-arm trials. RESULTS: To aid the wider implementation of multi-arm clinical trial designs, we have developed a web application for sample size calculation when using a variety of popular multiple comparison corrections. Furthermore, the application supports sample size calculation to control several varieties of power, as well as the determination of optimised arm-wise allocation ratios. It is built using the Shiny package in the R programming language, is free to access on any device with an internet browser, and requires no programming knowledge to use. It incorporates a variety of features to make it easier to use, including help boxes and warning messages. Using design parameters motivated by a recently completed phase II oncology trial, we demonstrate that the application can effectively determine and evaluate complex multi-arm trial designs. CONCLUSIONS: The application provides the core information required by statisticians and clinicians to review the operating characteristics of a chosen multi-arm clinical trial design. The range of designs supported by the application is broader than other currently available software solutions. Its primary limitation, particularly from a regulatory agency point of view, is its lack of validation. However, we present an approach to efficiently confirming its results via simulation.


Asunto(s)
Ensayos Clínicos como Asunto/métodos , Interpretación Estadística de Datos , Humanos , Proyectos de Investigación , Tamaño de la Muestra , Programas Informáticos , Navegador Web
SELECCIÓN DE REFERENCIAS
DETALLE DE LA BÚSQUEDA