Accurate and efficient P-values for rank-based independence tests with clustered data using a saddlepoint approximation

  • 0Department of Mathematics, Faculty of Education, Ain Shams University, Cairo, 11511, Egypt. haidynewer@edu.asu.edu.eg.

|

Summary

This summary is machine-generated.

Related Concept Videos

Introduction to Test of Independence 01:21

2.9K

In statistics, the term independence means that one can directly obtain the probability of any event involving both variables by multiplying their individual probabilities. Tests of independence are chi-square tests involving the use of a contingency table of observed (data) values.
The test statistic for a test of independence is similar to that of a goodness-of-fit test:

where:

O = observed values
E = expected values (which should be at least 5)

A test of independence determines whether...

Wilcoxon Rank-Sum Test 01:21

681

The Wilcoxon rank-sum test, also known as the Mann-Whitney U test, is a nonparametric test used to determine if there is a significant difference between the distributions of two independent samples. This test is designed specifically for two independent populations and has the following key requirements:

The samples must be randomly drawn.
The data should be ordinal or capable of being converted to an ordinal scale, allowing the values to be ordered and ranked.

The null hypothesis is that...

Friedman Two-way Analysis of Variance by Ranks 01:21

471

Friedman's Two-Way Analysis of Variance by Ranks is a nonparametric test designed to identify differences across multiple test attempts when traditional assumptions of normality and equal variances do not apply. Unlike conventional ANOVA, which requires normally distributed data with equal variances, Friedman's test is ideal for ordinal or non-normally distributed data, making it particularly useful for analyzing dependent samples, such as matched subjects over time or repeated measures...

Hypothesis Test for Test of Independence 01:16

7.4K

The test of independence is a chi-square-based test used to determine whether two variables or factors are independent or dependent. This hypothesis test is used to examine the independence of the variables. One can construct two qualitative survey questions or experiments based on the variables in a contingency table. The goal is to see if the two variables are unrelated (independent) or related (dependent). The null and alternative hypotheses for this test are:
H0: The two variables (factors)...

Expected Frequencies in Goodness-of-Fit Tests 01:19

7.1K

A goodness-of-fit test is conducted to determine whether the observed frequency values are statistically similar to the frequencies expected for the dataset. Suppose the expected frequencies for a dataset are equal such as when predicting the frequency of any number appearing when casting a die. In that case, the expected frequency is the ratio of the total number of observations (n)  to the number of categories (k).

Hence, the expected frequency of any number appearing when casting a die...

Fisher's Exact Test 01:08

1.1K

Fisher's exact test is a statistical significance test widely used to analyze 2x2 contingency tables, particularly in situations where sample sizes are small. Unlike the chi-squared test, which approximates P-values and assumes minimum expected frequencies of at least five in each cell, Fisher's exact test calculates the exact probability (P-value) of observing the data or more extreme results under the null hypothesis. This feature makes it especially valuable when the assumptions of...