Aggregating dependent signals with heavy-tailed combination tests
View abstract on PubMed
Summary
This summary is machine-generated.Combining p-values from multiple statistical tests is challenging. New methods using Cauchy and harmonic means show promise for dependent p-values, offering power gains over traditional tests in certain scenarios.
Area Of Science
- Statistical inference
- Dependence modeling
- Hypothesis testing
Background
- Combining dependent p-values is a significant challenge in statistical inference.
- Methods like Cauchy and harmonic mean p-value combination are gaining attention for their robustness to unknown dependence.
- Evaluating these methods under asymptotic regimes is crucial for understanding their behavior.
Purpose Of The Study
- To theoretically and empirically evaluate Cauchy and harmonic mean p-value combination tests.
- To investigate the performance of these tests under different types of p-value dependence.
- To compare their validity and power against the Bonferroni test as significance levels approach zero.
Main Methods
- Asymptotic analysis of p-value combination tests.
- Examination of pairwise asymptotically independent and quasi-asymptotically dependent p-values.
- Monte Carlo simulations to assess test validity and power.
- Analysis of test performance based on distribution support and tail heaviness.
Main Results
- Under pairwise asymptotic independence, combination tests are asymptotically valid but converge to the Bonferroni test as significance levels decrease.
- Under pairwise quasi-asymptotic dependence, simulations indicate these tests remain valid and offer power advantages over the Bonferroni test.
- Test performance is influenced by the support and tail heaviness of the underlying distributions.
Conclusions
- Cauchy and harmonic mean p-value combination tests show potential, especially when p-values exhibit substantial dependence.
- These methods can outperform the Bonferroni test in specific dependent scenarios.
- Further investigation into distribution properties is warranted for optimal test selection.
Related Concept Videos
A goodness-of-fit test is conducted to determine whether the observed frequency values are statistically similar to the frequencies expected for the dataset. Suppose the expected frequencies for a dataset are equal such as when predicting the frequency of any number appearing when casting a die. In that case, the expected frequency is the ratio of the total number of observations (n) to the number of categories (k).
Hence, the expected frequency of any number appearing when casting a die...
The sign test for matched pairs offers a robust method for comparing two paired samples, often for the effects of an intervention in one of them. This method is very useful in situations where the underlying distribution of the data is unknown. The test compares two related samples—often pre- and post-treatment measurements on the same subjects—to determine if there are significant differences in their median values.
To conduct the sign test, we first calculate the differences in...
Sometimes, a data set can have a recorded numerical observation that greatly deviates from the rest of the data. Assuming that the data is normally distributed, a statistical method called the Grubbs test can be used to determine whether the observation is truly an outlier. To perform a two-tailed Grubbs test, first, calculate the absolute difference between the outlier and the mean. Then, calculate the ratio between this difference and the standard deviation of the sample. This...
In statistics, the term independence means that one can directly obtain the probability of any event involving both variables by multiplying their individual probabilities. Tests of independence are chi-square tests involving the use of a contingency table of observed (data) values.
The test statistic for a test of independence is similar to that of a goodness-of-fit test:
where:
O = observed values
E = expected values (which should be at least 5)
A test of independence determines whether...
Significance testing is a set of statistical methods used to test whether a claim about a parameter is valid. In analytical chemistry, significance testing is used primarily to determine whether the difference between two values comes from determinate or random errors. The effect of a particular change in the measurement protocol, analyst, or sample itself can cause a deviation from the expected result. In the case of a suspected deviation/outlier, we need to be able to confirm mathematically...
Hypothesis testing is a critical statistical procedure facilitating informed, evidence-based decisions. It begins with a hypothesis, which is a tentative explanation, or a prediction about a population parameter. This hypothesis can be either a null hypothesis (H0), indicating no effect or difference, or an alternative hypothesis (Ha), suggesting an effect or difference.
Statistical significance measures the probability that an observed result occurred by chance. If this probability, known as...

