Statistical Hypothesis Testing with Python Implementation
September 16, 2021
Hypothesis testing is the analysis of assumptions on a population sample. In other words, it involves checking whether a hypothesis should be accepted or not.
Hypothesis testing has improved decision-making in different sectors including business. Today, organizations rely on hypothesis testing because of the enormous amount of data generated across the globe.
Table of contents
- Understanding hypothesis testing
- Confidence of null hypothesis
- Statistical techniques for hypothesis testing
- Chi-square test with Python implementation
- T-Test with Python implementation
- ANOVA with Python implementation
- Additional resources
At the end of this tutorial, readers should be able to:
- understand statistical hypothesis testing,
- perform t-test, chi-squared test & ANOVA using Python and a new dataset,
- differentiate between the two types of null hypothesis errors,
- explain the confidence of the null hypothesis,
- use Pandas functions for data analysis, and
- identify the best areas to apply t-test, chi-square test, and ANOVA test.
Understanding hypothesis testing
The two types of hypothesis testing are null hypothesis and alternate hypothesis.
Null hypothesis is the initial assumption about an event (also referred to as the ground truth).
Alternate hypothesis is an assumption that counters the initial assumption.
Consider a situation where a seventy-year-old woman has a visible bump in her belly. Medical officers may assume that the bump is fibroid.
In this case, our initial conclusion (or our null hypothesis) is that she has a fibroid while our alternate hypothesis is that she does not have a fibroid.
To carry out hypothesis testing, we will refer to the null hypothesis (initial assumption) as the H0 hypothesis and the alternate hypothesis (counter assumption) as the H1 hypothesis.
The follow-up action is to collect the available data samples to support the null hypothesis.
We should collect data pertaining to the hypothesis and analyze it to decide if H0 can be accepted or rejected.
While doing that, there is a likelihood of the following events happening:
The ground truth (H0) is true, so H0 is accepted.
The ground truth (H0) is not true, so H0 is rejected and H1 is accepted.
The above two cases are the desired possibilities. It’s either our null hypothesis was right and adopted or our null hypothesis was wrong and rejected.
The remaining possibilities are outlined below:
Null hypothesis (H0) is true but we reject it.
Null hypothesis (H0) is not true, but we did not reject it.
These cases are not desirable since we have not recognized the right hypothesis.
Type-1 error occurs when we reject the null hypothesis even though it was true.
Type-2 error involves accepting the null hypothesis even though it was wrong.
The severity of type-1 and type-2 errors depends on the task at hand. Data analysts should therefore determine the errors that could impact the outcome negatively.
Base on the example above, it is essential to be sure due to the huge risks involved. For instance, it can be considered that the patient does not have fibroids while in real life she is affected. It would be safer to conclude that the consequences for making a type-1 error, in this case, are much more grave than making a type-2 error.
When making assumptions, it’s crucial to use the p-value to determine whether to accept or dismiss the null hypothesis.
Confidence of null hypothesis
A p-value explains the likelihood of an assumption being true based on the null hypothesis. It is an abbreviation for probability value.
Technically, the only way we can accept or reject our null hypothesis is after determining our p-value.
The smaller our p-value is, the more delicate it is to trust our null hypothesis.
To calculate our p-value, there should be a certain level of significance to the initial assumption.
However, the p-value is usually within the range of 0 and 1. In most cases, the threshold value is set at 0.05 before the experiments in order to avoid bias.
We can also have a p-value greater than 0.05 e.g 0.1, the bigger or smaller the p-value is, depends on the task at hand, and how much confidence we have in the null hypothesis.
The alpha level allows us to ascertain whether to consider accepting or rejecting the null hypothesis. It also refers to the probability of making a bad choice if the ground truth (H0) is true.
The alpha level is determined by the following formula:
alpha = 1 - confidence level
The level of confidence we have in the null hypothesis is measured in a percentage, having 100% confidence is saying the statement is a fact, but once the confidence level dropped out of 100, it means there is a level of doubt to it and at that point, hypothesis testing is needed on it.
confidence level from what we’ve seen in our data is 95%, then the
alpha will be 0.05:
alpha = 1 - 0.95
It is important to establish the alpha level correctly to avoid biased decisions, especially when we dismiss the null hypothesis. The standard or preferred level is 5%.
Statistical techniques for hypothesis testing
There are three popular methods of hypothesis testing.
- Chi-square test
- ANOVA test
The chi-square test is adopted when there is a need to analyze two categorical elements in a data set. It focuses on the relationship between these two categorical variables. For example, comparing the after-effects of malaria drug A to those of malaria drug B.
Chi-square is a tool used when checking how divergent the observed frequency is from the expected results.
In a chi-square test, we acknowledge the null hypothesis when two variables, A and B are not dependent or have no relationship between them.
We call upon the alternate hypothesis if variable A depends on variable B or vice versa.
The formula for calculating the chi-square test is shown below:
Consider the following example:
A mathematics teacher conducted a test for a class of 20 students. He expected 5 to score good marks, 7 to score average marks while 8 to fail. Eventually, when the result came out, 5 failed, 6 were average while 9 scored good marks.
We can use a chi-square test to analyze the performance of this class.
From the table above, students fall into three categories ranging from what is expected to what is finally recorded. The main categories are
Let’s denote the expected and observed values with letters to make them more understandable.
observed good be
Aand the expected good be
observed average be
Band the expected average be
observed fail be
Cand the expected to fail to be
We then find the difference between the expected and observed values, as shown below:
For huge datasets, mathematical calculations cannot be done by hand. The implementation of the chi-square test in Python is discussed in the next section.
Chi-square test with Python implementation
We need to import
matplotlib, as demonstrated below:
import numpy as np import pandas as pd import scipy.stats as stats import matplotlib.pyplot as plt import statsmodels.api as sm from scipy.stats import f_oneway from scipy.stats import chi2_contingency from statsmodels.formula.api import ols import warnings warnings.filterwarnings( "ignore" )
The libraries imported are the most important libraries used for any data analysis project.
chi2_contigency libraries are used when implementing one-way ANOVA and chi-square tests respectively.
In this tutorial, we will be using the popular Iris dataset. It can be downloaded from here.
The next phase is to load the data into the notebook:
On reading the dataset into the notebook, it is important to add columns name to it for a better understanding of the data we are working with. Hence, we proceed to check the first five to confirm our formatting works.
df = pd.read_csv("/content/drive/MyDrive/IRIS.csv") df.columns = ['sepal_length', 'sepal_width', 'petal_length', 'petal_width', 'species'] df.head()
df.sample() to visualize samples randomly from our dataframe.
Always check for missing values to avoid errors and be sure there is consistency going forward.
sepal_length 0 sepal_width 0 petal_length 0 petal_width 0 species 0 dtype: int64
Confirm the categories in the target class.
array(['Iris-setosa', 'Iris-versicolor', 'Iris-virginica'], dtype=object)
We can be sure now that there are 3 classes of species and there are no missing values.
Let’s proceed with the data manipulation.
petal_width to compare species for our chi-square test.
The first move is to check the summary of
petal_width. We then use this information to transform it into a categorical variable.
count 150.000000 mean 1.198667 std 0.763161 min 0.100000 25% 0.300000 50% 1.300000 75% 1.800000 max 2.500000 Name: petal_width, dtype: float64
The chi-square test helps in determining if there is a notable difference between observed and normal frequencies in one or more categories.
The values of
petal_width must be changed into categories of zeros and ones using the 50% percentile.
There are two conditions that must be satisfied while calculating percentile:
Data must be arranged in ascending order.
The index must be rounded up if it is not an integer.
A quartile is another term relating to percentile. You can find out more about percentiles and quartiles here.
Implementing a function to create the new petal width into the dataframe with the percentile calculated.
def petal_cat(df): if df['petal_width'] <= 1.3: return 0 elif df['petal_width'] > 1.3: return 1 else: return 'Indifferent' df['petal_width_new'] = df.apply(petal_cat, axis=1)
The data format for the
species class has to be transformed to 0, 1, and 2, which is a machine-readable format to facilitate testing.
Therefore, with label-encoding, a numeric value representation will be given to each of the categories in the class.
To learn more about label encoding, the reader is advised to check up on the following resource.
def species_cat(df): if df["species"] == "Iris-virginica": return 0 elif df["species"] == "Iris-versicolor": return 1 else: return 2 df["species"] = df.apply(species_cat, axis=1)
Confirm the function works perfectly by printing randomly five samples.
Drop all other columns except the two columns that we need, the
species and the
df_new = df.drop(columns=["sepal_width", "sepal_length", "petal_length", "petal_width"]) print(df_new.head())
Let’s make use of the
chic2_contigency library that we had imported into the project.
stat, p, dof, expected = chi2_contingency(df_new) print("The degree of freedom is: ", dof)
The degree of freedom is: 149
The three outputs in test statistics will include the degree of freedom, p-value, and expected values.
Test statistic is a characterized feature of significance in the chi-square test. It helps in determining how uncommon the result might be, provided that the null hypothesis is agreed to be true.
The number of features or variables that we have access to varies depending on the degree of freedom.
The probability of confidence is similar to the level of confidence explained earlier, it is about how close our assumption is inclined towards being a fact, it is the amount of confidence we have in our null hypothesis, it is therefore set to 95%.
prob = 0.95 alpha = 1.0 - prob print('The alpha/significance level = %.3f' % alpha) print('The p-value is = %.2f' % p) if p <= alpha: print('Reject the Null Hypothesis (Reject H0)') else: print('Accept the Null Hypothesis (Do not reject H0)')
The alpha/significance level = 0.050 The p-value is = 0.10 Accept the Null Hypothesis (Do not reject H0)
In the results above, the p-value surpasses the alpha value set at 0.05. We will acknowledge the null hypothesis and as well dismiss the alternate hypothesis.
This means the petal width and species of flower are not dependent on each other, i.e. there is no connection between them.
T-Test with Python implementation
The t-test is utilized when we plan to evaluate the discrepancy between the means of two groups of samples. Unlike the chi-square test, the t-test is used on continuous variables.
T-test has three basic types.
- One sample t-test: It checks whether the population has a different mean from the sample.
The mathematical expression is:
the observed mean group is
assumed population means is
the standard deviation of the data group is
the number of observations in the group is
- Two-sample t-test: It checks and juxtaposes the means of two groups that are not dependent on each other and compares the population’s means to ascertain if there is a huge difference.
The mathematical expression is:
Where, data sample A’s mean is
Ma, data sample B’s mean is
Mb, the size of sample A is
Na, the size of sample B is
Nb, and variance is
- Paired t-test: It is used in checking and comparing the means of different samples from a group.
Mathematically, it can be resolved with:
Where, the difference between paired observations’ sample mean is
D, the assumed mean difference is
m, standard deviation is
xd, and the number of observations in the group is
Performing a one-sample T-test
For the sake of a one-sample t-test, we will be re-assigning the column of petal width into a single dataframe for ease of use.
df_ = df.petal_width
The pandas dataframe
sample() function is used here as well to have a random view of our dataframe.
103 1.8 131 2.0 114 2.4 35 0.2 38 0.2 Name: petal_width, dtype: float64
Applying the one sample t-test function of the dataframe.
H0: The mean of petal_width is 1.199
H1: The mean of petal_width is not 1.199
Analyzing the one-sample t-test, it finds out if the hypothesized mean is similar to or different from the group’s mean. From the example above, we’ve selected the
petal_width as the population sample to perform the test on.
From the test, since the p-value beats the alpha level set at 0.05, we acknowledge the null hypothesis because we don’t have enough evidence to prove otherwise.
Moreover, if you check, the p-value is very close to 1.0, which means that the mean of
petal_width is most likely 1.199.
The evidence is too strong to be rejected, it is almost accurate.
Performing a two-sample T-test
The standard for the two-sample t-test is for the two independent groups we are sampling to have equal variances. We can only know the variances are the same when the ratio of the higher to the lower variance is less than 4:1.
Checking the variance is only making sure that two populations are probably spread out or normally distributed, this will help prevent bias.
Hence, the null hypothesis will be assuming that since the two groups are normally distributed, they likely should have the same mean.
class1 = df.petal_width class2 = df.sepal_length
Find variance for each group.
Do the ratio check to know if it satisfies the condition for the two-sample t-test. The ratio check is done on the variance to ensure that both of the populations are normally distributed.
ratio_check = np.var(class2)/np.var(class1) print(ratio_check)
Obviously, the ratio is less than 4:1, thus the variances are considered to be equal.
Applying the two-sample t-test function on our classes,
print(stats.ttest_ind(a=class1, b=class2, equal_var=True))
H0: The mean of both samples (petal_width and sepal_length) are equal.
H1: The mean of both samples (petal_width and sepal_length) are not equal.
p-value (p=3.7974378831185126e-148) is far less than
alpha = 0.05.
However, interpreting any form of t-test largely depends on the
p-value, the result above explains that the means of the two samples (
sepal_length) are different.
Therefore we will dismiss the null hypothesis which says the mean of the two samples are the same because their variances are assumed to be equal.
There is no evidence sustainable enough to prove the two populations have the same mean. H1 (alternate hypothesis) is however true.
ANOVA with Python implementation
ANOVA is a word coined from ‘Analysis of Variance’. It is a statistical concept that shows the differences between the means of more than two independent groups, using variance analysis on samples from those groups.
It is used for checking the contrast between three or more samples with one test. Especially when the categorical class has over two categories.
With ANOVA, you get to discover obvious differences between the means of your independent features. Upon getting a clearer picture of the differences, you can understand how each of them connects to your dependent variable. You can see what are the influencing factors for the relationship.
A generic example could be as follows:
The growth department of a certain company discovered an increase in sales for a particular product. In recent times, they’ve tried several advertisement channels so it is difficult to say which one yielded the sales increase the most.
They did online posters, billboards, TV ads, mouth-to-mouth, and sponsorship. By carrying out an ANOVA test on the several advertisement channels, they will be able to better understand which one works best out of all channels (independent variables), with the connection it has with the increase in sales which is the dependent variable.
During ANOVA testing, the hypothesis is:
H0: When all samples’ means are the same.
H1: When one or more samples are very much different.
One way ANOVA test
This is employed to determine the effect of a variable on one or two other variables by comparing their means. Using the same petal example we will check if
petal width, has an effect on
petal length and
sepal length using one-way ANOVA test.
class1 = df.petal_width class2 = df.sepal_length class3 = df.petal_length print(f_oneway(class1, class2, class3))
p-value (5.738282463819433e-122) is far less than
alpha (0.05), we dismiss the null hypothesis, as there exists no evidence sustainable enough to accept it.
This means that the sample means are very different. Meaning that our H1 (alternate hypothesis) is true.
Two way ANOVA test
This is called when we are dealing with three or more variables, trying to compare their means with each other.
petal = df.petal_width sepal = df.sepal_length species = df.species model = ols('species ~ C(petal) + C(sepal) + C(petal):C(sepal)', data=df).fit() print(sm.stats.anova_lm(model, typ=2))
The p-value of
petal_width is more than 0.05, which increases the probability of our null hypothesis being accepted.
The p-value of
sepal_length is less than 0.05, which increases the probability of our null hypothesis being rejected.
Likewise, the p-value of both
sepal_length is lower than 0.05, we dismiss the null hypothesis since it is obvious one or more samples are very much different in their means. That is the same thing as saying, the average of all data samples is not equal.
At the end of this tutorial, I believe the readers should have a better grasp on the concept of statistical hypothesis testing and how to implement it with Python.
To carry out hypothesis testing, the first step is to form an initial assumption and label it as H0. The next step would be to collect all data samples available to support our hypothesis, collect all the shreds of evidence and analyze the data, and make a decision whether to accept the H0 or reject it.
When we reject the ground truth but it is true, we encounter the Type 1 error. On the flip side, when we do not reject the ground truth despite the fact that it is not true, we encounter the Type 2 error.
Additional resources are added below to better solidify the knowledge gained from this tutorial.
- Test of Hypothesis in Data Science
- ANOVA Test: Definition, Types, Examples
- Understanding t-Tests: t-values and t-distributions
- Make sense with two sample t-test
- Understanding Hypothesis Testing: Level of Significance (Alpha) and P values in Statistics
- Chi-squared test application
- Chi-square mathematics for statistics students.
Peer Review Contributions by: Prashanth Saravanan