Module stikpetP.tests.test_score_os
Expand source code
import pandas as pd
from statistics import NormalDist
def ts_score_os(data, p0 = 0.5, p0Cat=None, codes = None, cc = None):
'''
One-sample Score Test
---------------------
A one-sample score test could be used with binary data, to test if the two categories have a significantly different proportion. It is an approximation of a binomial test, by using a standard normal distribution. Since the binomial distribution is discrete while the normal is continuous, a so-called continuity correction can (should?) be applied.
The null hypothesis is usually that the proportions of the two categories in the population are equal (i.e. 0.5 for each). If the p-value of the test is below the pre-defined alpha level (usually 5% = 0.05) the null hypothesis is rejected and the two categories differ in proportion significantly.
The input for the function doesn't have to be a binary variable. A nominal variable can also be used and the two categories to compare indicated. A significance in general is the probability of a result as in the sample, or more extreme, if the null hypothesis is true.
This function is shown in this [YouTube video](https://youtu.be/06q7qlTOs-s) and the test is also described at [PeterStatistics.com](https://peterstatistics.com/Terms/Tests/proportion-one-sample.html)
Parameters
----------
data : list or pandas data series
the data
p0 : float, optional
hypothesized proportion for the first category (default is 0.5)
p0Cat : optional
the category for which p0 was used
codes : list, optional
the two codes to use
cc : {None, "yates"}, optional
use continuity correction. Default is None
Returns
-------
pandas.DataFrame
A dataframe with the following columns:
- *n* : the sample size
- *statistic* : test statistic (z-value)
- *p-value (2-sided)* : two-sided significance (p-value)
- *test* : test used
Notes
-----
Also sometimes called a 'proportion' test.
To decide on which category is associated with p0 the following is used:
* If codes are provided, the first code is assumed to be the category for the p0.
* If p0Cat is specified that will be used for p0 and all other categories will be considered as category 2, this means if there are more than two categories the remaining two or more (besides p0Cat) will be merged as one large category.
* If neither codes or p0Cat is specified and more than two categories are in the data a warning is printed and no results.
* If neither codes or p0Cat is specified and there are two categories, p0 is assumed to be for the category closest matching the p0 value (i.e. if p0 is above 0.5 the category with the highest count is assumed to be used for p0)
The formula used is (Wilson, 1927):
$$z=\\frac{x - \\mu}{SE}$$
With:
$$\\mu = n\\times p_0$$
$$SE=\\sqrt{\\mu\\times\\left(1-p_0\\right)}$$
Symbols used:
* $x$ is the number of successes in the sample
* $p_0$ the expected proportion (i.e. the proportion according to the null hypothesis)
If the Yates continuity correction is used the formula changes to (Yates, 1934, p. 222):
$$z_{Yates} = \\frac{\\left|x - \\mu\\right| - 0.5}{SE}$$
The formula used and naming comes from IBM (2021, p. 997) who refer to Agresti, most likeli Agresti (2013, p. 10)
It uses NormalDist from Python's statistics library
Before, After and Alternatives
------------------------------
Before running the test you might first want to get an impression using a frequency table:
[tab_frequency](../other/table_frequency.html#tab_frequency)
After the test you might want an effect size measure:
* [es_cohen_g](../effect_sizes/eff_size_cohen_g.html#es_cohen_g) for Cohen g
* [es_cohen_h_os](../effect_sizes/eff_size_cohen_h_os.html#es_cohen_h_os) for Cohen h'
* [es_alt_ratio](../effect_sizes/eff_size_alt_ratio.html#es_alt_ratio) for Alternative Ratio
* [r_rosenthal](../correlations/cor_rosenthal.html#r_rosenthal) for Rosenthal Correlation
Alternatives for this test could be:
* [ts_binomial_os](../tests/test_binomial_os.html#ts_binomial_os) for One-Sample Binomial Test
* [ts_wald_os](../tests/test_wald_os.html#ts_wald_os) for One-Sample Wald Test
References
----------
Agresti, A. (2013). *Categorical data analysis* (3rd ed.). Wiley.
IBM SPSS Statistics Algorithms. (2021). IBM.
Wilson, E. B. (1927). Probable Inference, the Law of Succession, and Statistical Inference. *Journal of the American Statistical Association, 22*(158), 209–212. doi:10.2307/2276774
Yates, F. (1934). Contingency tables involving small numbers and the chi square test. *Supplement to the Journal of the Royal Statistical Society, 1*(2), 217–235. doi:10.2307/2983604
Author
------
Made by P. Stikker
Companion website: https://PeterStatistics.com
YouTube channel: https://www.youtube.com/stikpet
Donations: https://www.patreon.com/bePatron?u=19398076
Examples
---------
>>> pd.set_option('display.width',1000)
>>> pd.set_option('display.max_columns', 1000)
Example 1: Numeric list
>>> ex1 = [1, 1, 2, 1, 2, 1, 2, 1]
>>> ts_score_os(ex1)
n statistic p-value (2-sided) test
0 8 -0.707107 0.4795 one-sample score (assuming p0 for 1)
Setting a different hypothesized proportion, and going over different methods to determine two-sided test:
>>> ts_score_os(ex1, p0=0.3)
n statistic p-value (2-sided) test
0 8 -2.005944 0.044862 one-sample score (assuming p0 for 1)
>>> ts_score_os(ex1, p0=0.3, cc="yates")
n statistic p-value (2-sided) test
0 8 -1.620185 0.105193 one-sample score with Yates continuity correction (assuming p0 for 1)
Example 2: pandas Series
>>> df1 = pd.read_csv('https://peterstatistics.com/Packages/ExampleData/GSS2012a.csv', sep=',', low_memory=False, storage_options={'User-Agent': 'Mozilla/5.0'})
>>> ts_score_os(df1['sex'])
n statistic p-value (2-sided) test
0 1974 -4.546506 0.000005 one-sample score (assuming p0 for FEMALE)
>>> ts_score_os(df1['mar1'], codes=["DIVORCED", "NEVER MARRIED"])
n statistic p-value (2-sided) test
0 709 -3.042019 0.00235 one-sample score (with p0 for DIVORCED)
'''
if type(data) is list:
data = pd.Series(data)
#remove missing values
data = data.dropna()
#Determine number of successes, failures, and total sample size
if codes is None:
#create a frequency table
freq = data.value_counts()
if p0Cat is None:
#check if there were exactly two categories or not
if len(freq) != 2:
# unable to determine which category p0 would belong to, so print warning and end
print("WARNING: data does not have two unique categories, please specify two categories using codes parameter")
return
else:
#simply select the two categories as cat1 and cat2
n1 = freq.values[0]
n2 = freq.values[1]
n = n1 + n2
#determine p0 was for which category
p0_cat = freq.index[0]
if p0 > 0.5 and n1 < n2:
n3=n2
n2 = n1
n1 = n3
p0_cat = freq.index[1]
cat_used = " (assuming p0 for " + str(p0_cat) + ")"
else:
n = sum(freq.values)
n1 = sum(data==p0Cat)
n2 = n - n1
p0_cat = p0Cat
cat_used = " (with p0 for " + str(p0Cat) + ")"
else:
n1 = sum(data==codes[0])
n2 = sum(data==codes[1])
n = n1 + n2
cat_used = " (with p0 for " + str(codes[0]) + ")"
minCount = n1
ExpProp = p0
if (n2 < n1):
minCount = n2
ExpProp = 1 - ExpProp
#Normal approximation
if (cc is None):
p = minCount / n
q = 1 - p
se = (p0 * (1 - p0) / n)**0.5
Z = (p - ExpProp) / se
sig2 = 2 * (1 - NormalDist().cdf(abs(Z)))
testValue = Z
testUsed = "one-sample score"
elif not (cc is None) and (cc == "yates"):
#Normal approximation with continuity correction
p = (minCount + 0.5) / n
q = 1 - p
se = (p0 * (1 - p0) / n)**0.5
Z = (p - ExpProp) / se
sig2 = 2 * (1 - NormalDist().cdf(abs(Z)))
testValue = Z
testUsed = "one-sample score with Yates continuity correction"
testUsed = testUsed + cat_used
testResults = pd.DataFrame([[n, testValue, sig2, testUsed]], columns=["n", "statistic", "p-value (2-sided)", "test"])
return (testResults)
Functions
def ts_score_os(data, p0=0.5, p0Cat=None, codes=None, cc=None)
-
One-sample Score Test
A one-sample score test could be used with binary data, to test if the two categories have a significantly different proportion. It is an approximation of a binomial test, by using a standard normal distribution. Since the binomial distribution is discrete while the normal is continuous, a so-called continuity correction can (should?) be applied.
The null hypothesis is usually that the proportions of the two categories in the population are equal (i.e. 0.5 for each). If the p-value of the test is below the pre-defined alpha level (usually 5% = 0.05) the null hypothesis is rejected and the two categories differ in proportion significantly.
The input for the function doesn't have to be a binary variable. A nominal variable can also be used and the two categories to compare indicated. A significance in general is the probability of a result as in the sample, or more extreme, if the null hypothesis is true.
This function is shown in this YouTube video and the test is also described at PeterStatistics.com
Parameters
data
:list
orpandas data series
- the data
p0
:float
, optional- hypothesized proportion for the first category (default is 0.5)
p0Cat
:optional
- the category for which p0 was used
codes
:list
, optional- the two codes to use
cc
:{None, "yates"}
, optional- use continuity correction. Default is None
Returns
pandas.DataFrame
-
A dataframe with the following columns:
- n : the sample size
- statistic : test statistic (z-value)
- p-value (2-sided) : two-sided significance (p-value)
- test : test used
Notes
Also sometimes called a 'proportion' test.
To decide on which category is associated with p0 the following is used: * If codes are provided, the first code is assumed to be the category for the p0. * If p0Cat is specified that will be used for p0 and all other categories will be considered as category 2, this means if there are more than two categories the remaining two or more (besides p0Cat) will be merged as one large category. * If neither codes or p0Cat is specified and more than two categories are in the data a warning is printed and no results. * If neither codes or p0Cat is specified and there are two categories, p0 is assumed to be for the category closest matching the p0 value (i.e. if p0 is above 0.5 the category with the highest count is assumed to be used for p0)
The formula used is (Wilson, 1927): z=\frac{x - \mu}{SE}
With: \mu = n\times p_0 SE=\sqrt{\mu\times\left(1-p_0\right)}
Symbols used:
- $x$ is the number of successes in the sample
- $p_0$ the expected proportion (i.e. the proportion according to the null hypothesis)
If the Yates continuity correction is used the formula changes to (Yates, 1934, p. 222): z_{Yates} = \frac{\left|x - \mu\right| - 0.5}{SE}
The formula used and naming comes from IBM (2021, p. 997) who refer to Agresti, most likeli Agresti (2013, p. 10)
It uses NormalDist from Python's statistics library
Before, After and Alternatives
Before running the test you might first want to get an impression using a frequency table: tab_frequency
After the test you might want an effect size measure: * es_cohen_g for Cohen g * es_cohen_h_os for Cohen h' * es_alt_ratio for Alternative Ratio * r_rosenthal for Rosenthal Correlation
Alternatives for this test could be: * ts_binomial_os for One-Sample Binomial Test * ts_wald_os for One-Sample Wald Test
References
Agresti, A. (2013). Categorical data analysis (3rd ed.). Wiley.
IBM SPSS Statistics Algorithms. (2021). IBM.
Wilson, E. B. (1927). Probable Inference, the Law of Succession, and Statistical Inference. Journal of the American Statistical Association, 22(158), 209–212. doi:10.2307/2276774
Yates, F. (1934). Contingency tables involving small numbers and the chi square test. Supplement to the Journal of the Royal Statistical Society, 1(2), 217–235. doi:10.2307/2983604
Author
Made by P. Stikker
Companion website: https://PeterStatistics.com
YouTube channel: https://www.youtube.com/stikpet
Donations: https://www.patreon.com/bePatron?u=19398076Examples
>>> pd.set_option('display.width',1000) >>> pd.set_option('display.max_columns', 1000)
Example 1: Numeric list
>>> ex1 = [1, 1, 2, 1, 2, 1, 2, 1] >>> ts_score_os(ex1) n statistic p-value (2-sided) test 0 8 -0.707107 0.4795 one-sample score (assuming p0 for 1)
Setting a different hypothesized proportion, and going over different methods to determine two-sided test:
>>> ts_score_os(ex1, p0=0.3) n statistic p-value (2-sided) test 0 8 -2.005944 0.044862 one-sample score (assuming p0 for 1) >>> ts_score_os(ex1, p0=0.3, cc="yates") n statistic p-value (2-sided) test 0 8 -1.620185 0.105193 one-sample score with Yates continuity correction (assuming p0 for 1)
Example 2: pandas Series
>>> df1 = pd.read_csv('https://peterstatistics.com/Packages/ExampleData/GSS2012a.csv', sep=',', low_memory=False, storage_options={'User-Agent': 'Mozilla/5.0'}) >>> ts_score_os(df1['sex']) n statistic p-value (2-sided) test 0 1974 -4.546506 0.000005 one-sample score (assuming p0 for FEMALE) >>> ts_score_os(df1['mar1'], codes=["DIVORCED", "NEVER MARRIED"]) n statistic p-value (2-sided) test 0 709 -3.042019 0.00235 one-sample score (with p0 for DIVORCED)
Expand source code
def ts_score_os(data, p0 = 0.5, p0Cat=None, codes = None, cc = None): ''' One-sample Score Test --------------------- A one-sample score test could be used with binary data, to test if the two categories have a significantly different proportion. It is an approximation of a binomial test, by using a standard normal distribution. Since the binomial distribution is discrete while the normal is continuous, a so-called continuity correction can (should?) be applied. The null hypothesis is usually that the proportions of the two categories in the population are equal (i.e. 0.5 for each). If the p-value of the test is below the pre-defined alpha level (usually 5% = 0.05) the null hypothesis is rejected and the two categories differ in proportion significantly. The input for the function doesn't have to be a binary variable. A nominal variable can also be used and the two categories to compare indicated. A significance in general is the probability of a result as in the sample, or more extreme, if the null hypothesis is true. This function is shown in this [YouTube video](https://youtu.be/06q7qlTOs-s) and the test is also described at [PeterStatistics.com](https://peterstatistics.com/Terms/Tests/proportion-one-sample.html) Parameters ---------- data : list or pandas data series the data p0 : float, optional hypothesized proportion for the first category (default is 0.5) p0Cat : optional the category for which p0 was used codes : list, optional the two codes to use cc : {None, "yates"}, optional use continuity correction. Default is None Returns ------- pandas.DataFrame A dataframe with the following columns: - *n* : the sample size - *statistic* : test statistic (z-value) - *p-value (2-sided)* : two-sided significance (p-value) - *test* : test used Notes ----- Also sometimes called a 'proportion' test. To decide on which category is associated with p0 the following is used: * If codes are provided, the first code is assumed to be the category for the p0. * If p0Cat is specified that will be used for p0 and all other categories will be considered as category 2, this means if there are more than two categories the remaining two or more (besides p0Cat) will be merged as one large category. * If neither codes or p0Cat is specified and more than two categories are in the data a warning is printed and no results. * If neither codes or p0Cat is specified and there are two categories, p0 is assumed to be for the category closest matching the p0 value (i.e. if p0 is above 0.5 the category with the highest count is assumed to be used for p0) The formula used is (Wilson, 1927): $$z=\\frac{x - \\mu}{SE}$$ With: $$\\mu = n\\times p_0$$ $$SE=\\sqrt{\\mu\\times\\left(1-p_0\\right)}$$ Symbols used: * $x$ is the number of successes in the sample * $p_0$ the expected proportion (i.e. the proportion according to the null hypothesis) If the Yates continuity correction is used the formula changes to (Yates, 1934, p. 222): $$z_{Yates} = \\frac{\\left|x - \\mu\\right| - 0.5}{SE}$$ The formula used and naming comes from IBM (2021, p. 997) who refer to Agresti, most likeli Agresti (2013, p. 10) It uses NormalDist from Python's statistics library Before, After and Alternatives ------------------------------ Before running the test you might first want to get an impression using a frequency table: [tab_frequency](../other/table_frequency.html#tab_frequency) After the test you might want an effect size measure: * [es_cohen_g](../effect_sizes/eff_size_cohen_g.html#es_cohen_g) for Cohen g * [es_cohen_h_os](../effect_sizes/eff_size_cohen_h_os.html#es_cohen_h_os) for Cohen h' * [es_alt_ratio](../effect_sizes/eff_size_alt_ratio.html#es_alt_ratio) for Alternative Ratio * [r_rosenthal](../correlations/cor_rosenthal.html#r_rosenthal) for Rosenthal Correlation Alternatives for this test could be: * [ts_binomial_os](../tests/test_binomial_os.html#ts_binomial_os) for One-Sample Binomial Test * [ts_wald_os](../tests/test_wald_os.html#ts_wald_os) for One-Sample Wald Test References ---------- Agresti, A. (2013). *Categorical data analysis* (3rd ed.). Wiley. IBM SPSS Statistics Algorithms. (2021). IBM. Wilson, E. B. (1927). Probable Inference, the Law of Succession, and Statistical Inference. *Journal of the American Statistical Association, 22*(158), 209–212. doi:10.2307/2276774 Yates, F. (1934). Contingency tables involving small numbers and the chi square test. *Supplement to the Journal of the Royal Statistical Society, 1*(2), 217–235. doi:10.2307/2983604 Author ------ Made by P. Stikker Companion website: https://PeterStatistics.com YouTube channel: https://www.youtube.com/stikpet Donations: https://www.patreon.com/bePatron?u=19398076 Examples --------- >>> pd.set_option('display.width',1000) >>> pd.set_option('display.max_columns', 1000) Example 1: Numeric list >>> ex1 = [1, 1, 2, 1, 2, 1, 2, 1] >>> ts_score_os(ex1) n statistic p-value (2-sided) test 0 8 -0.707107 0.4795 one-sample score (assuming p0 for 1) Setting a different hypothesized proportion, and going over different methods to determine two-sided test: >>> ts_score_os(ex1, p0=0.3) n statistic p-value (2-sided) test 0 8 -2.005944 0.044862 one-sample score (assuming p0 for 1) >>> ts_score_os(ex1, p0=0.3, cc="yates") n statistic p-value (2-sided) test 0 8 -1.620185 0.105193 one-sample score with Yates continuity correction (assuming p0 for 1) Example 2: pandas Series >>> df1 = pd.read_csv('https://peterstatistics.com/Packages/ExampleData/GSS2012a.csv', sep=',', low_memory=False, storage_options={'User-Agent': 'Mozilla/5.0'}) >>> ts_score_os(df1['sex']) n statistic p-value (2-sided) test 0 1974 -4.546506 0.000005 one-sample score (assuming p0 for FEMALE) >>> ts_score_os(df1['mar1'], codes=["DIVORCED", "NEVER MARRIED"]) n statistic p-value (2-sided) test 0 709 -3.042019 0.00235 one-sample score (with p0 for DIVORCED) ''' if type(data) is list: data = pd.Series(data) #remove missing values data = data.dropna() #Determine number of successes, failures, and total sample size if codes is None: #create a frequency table freq = data.value_counts() if p0Cat is None: #check if there were exactly two categories or not if len(freq) != 2: # unable to determine which category p0 would belong to, so print warning and end print("WARNING: data does not have two unique categories, please specify two categories using codes parameter") return else: #simply select the two categories as cat1 and cat2 n1 = freq.values[0] n2 = freq.values[1] n = n1 + n2 #determine p0 was for which category p0_cat = freq.index[0] if p0 > 0.5 and n1 < n2: n3=n2 n2 = n1 n1 = n3 p0_cat = freq.index[1] cat_used = " (assuming p0 for " + str(p0_cat) + ")" else: n = sum(freq.values) n1 = sum(data==p0Cat) n2 = n - n1 p0_cat = p0Cat cat_used = " (with p0 for " + str(p0Cat) + ")" else: n1 = sum(data==codes[0]) n2 = sum(data==codes[1]) n = n1 + n2 cat_used = " (with p0 for " + str(codes[0]) + ")" minCount = n1 ExpProp = p0 if (n2 < n1): minCount = n2 ExpProp = 1 - ExpProp #Normal approximation if (cc is None): p = minCount / n q = 1 - p se = (p0 * (1 - p0) / n)**0.5 Z = (p - ExpProp) / se sig2 = 2 * (1 - NormalDist().cdf(abs(Z))) testValue = Z testUsed = "one-sample score" elif not (cc is None) and (cc == "yates"): #Normal approximation with continuity correction p = (minCount + 0.5) / n q = 1 - p se = (p0 * (1 - p0) / n)**0.5 Z = (p - ExpProp) / se sig2 = 2 * (1 - NormalDist().cdf(abs(Z))) testValue = Z testUsed = "one-sample score with Yates continuity correction" testUsed = testUsed + cat_used testResults = pd.DataFrame([[n, testValue, sig2, testUsed]], columns=["n", "statistic", "p-value (2-sided)", "test"]) return (testResults)