Accurate and efficient P-values for rank-based independence tests with clustered data using a saddlepoint approximation
View abstract on PubMed
Summary
This summary is machine-generated.Accurate statistical inference for clustered data is challenging. A new double saddlepoint approximation method provides precise p-values and confidence intervals, overcoming computational limits of permutation tests for rank-based analyses.
Area Of Science
- Biostatistics
- Statistical Inference
- Clinical Trials
Background
- Clustered data in multi-center trials and longitudinal studies present statistical inference challenges due to within-cluster correlation.
- Rank-based tests (logrank, Wilcoxon) are robust but can have inflated Type I errors with standard approximations.
- Exact permutation tests are accurate but computationally infeasible for large datasets.
Purpose Of The Study
- To develop a computationally efficient and accurate statistical method for analyzing clustered data.
- To provide reliable p-values and confidence intervals for rank-based tests in the presence of within-cluster correlation.
- To address the methodological gap between the accuracy of permutation tests and the practicality of asymptotic methods.
Main Methods
- A novel double saddlepoint approximation framework is proposed.
- The method utilizes a permutation distribution reformulation via block urn design to maintain cluster integrity.
- This reformulation allows the test statistic's distribution to be represented as a sum of independent conditional random variables for saddlepoint computation.
Main Results
- The double saddlepoint method accurately controls Type I error rates for rank-based tests on clustered data.
- Performance is statistically equivalent to exact permutation tests but with significantly reduced computational cost.
- The approach handles right-censored survival data and tied ranks effectively.
Conclusions
- The proposed double saddlepoint approximation offers a practical, efficient, and statistically rigorous tool for biostatisticians analyzing clustered data.
- It overcomes the computational limitations of permutation tests while maintaining accuracy.
- The method can prevent false-positive conclusions sometimes arising from standard asymptotic approximations in clinical trial data.
Related Concept Videos
In statistics, the term independence means that one can directly obtain the probability of any event involving both variables by multiplying their individual probabilities. Tests of independence are chi-square tests involving the use of a contingency table of observed (data) values.
The test statistic for a test of independence is similar to that of a goodness-of-fit test:
where:
O = observed values
E = expected values (which should be at least 5)
A test of independence determines whether...
The Wilcoxon rank-sum test, also known as the Mann-Whitney U test, is a nonparametric test used to determine if there is a significant difference between the distributions of two independent samples. This test is designed specifically for two independent populations and has the following key requirements:
The samples must be randomly drawn.
The data should be ordinal or capable of being converted to an ordinal scale, allowing the values to be ordered and ranked.
The null hypothesis is that...
Friedman's Two-Way Analysis of Variance by Ranks is a nonparametric test designed to identify differences across multiple test attempts when traditional assumptions of normality and equal variances do not apply. Unlike conventional ANOVA, which requires normally distributed data with equal variances, Friedman's test is ideal for ordinal or non-normally distributed data, making it particularly useful for analyzing dependent samples, such as matched subjects over time or repeated measures...
The test of independence is a chi-square-based test used to determine whether two variables or factors are independent or dependent. This hypothesis test is used to examine the independence of the variables. One can construct two qualitative survey questions or experiments based on the variables in a contingency table. The goal is to see if the two variables are unrelated (independent) or related (dependent). The null and alternative hypotheses for this test are:
H0: The two variables (factors)...
A goodness-of-fit test is conducted to determine whether the observed frequency values are statistically similar to the frequencies expected for the dataset. Suppose the expected frequencies for a dataset are equal such as when predicting the frequency of any number appearing when casting a die. In that case, the expected frequency is the ratio of the total number of observations (n) to the number of categories (k).
Hence, the expected frequency of any number appearing when casting a die...
Fisher's exact test is a statistical significance test widely used to analyze 2x2 contingency tables, particularly in situations where sample sizes are small. Unlike the chi-squared test, which approximates P-values and assumes minimum expected frequencies of at least five in each cell, Fisher's exact test calculates the exact probability (P-value) of observing the data or more extreme results under the null hypothesis. This feature makes it especially valuable when the assumptions of...

