Understand Hypothesis testing with real-life examples

hypothesis-testing-real-life-example

If you are learning or working in the field of Data Science, you should have clear understanding of what are hypothesis testing and p values in statistics. If you learn those two things then understanding any basic machine learning algorithms like linear regression, and logistic regression will be easier for you. In this tutorial I will explain hypothesis testing with real life examples.

Many people have lots of confusion about hypothesis testing and p value. In this tutorial, I will explain you what is hypothesis testing in statistics with a very simple and real life example.

What is Hypothesis

Hypothesis testing is a statistical method to evaluate or investigate the validity of a claim or assumption about something based on sample data.

In simple English, meaning of Hypothesis is guess. Now a guess can be anything like: With a ladder tall enough, I could touch the sky or If I had wings, I could fly to the moon, If I had a time machine, I could travel to any point in history, etc.

Some guesses are just imagination, you can not prove or investigate them with anything. If we can investigate or test any guesses with data, that test is called Hypothesis testing in statistics.

Types of Hypothesis Testing

At this point, we know what is hypothesis testing and its use. Now there are two types of hypotheses:

  1. Null Hypothesis (H0)
  2. Alternative Hypothesis (H1)

Null Hypothesis

In general programming language, we denote Null as empty or zero. In statistics, Null hypothesis means default hypothesis, the hypothesis or guess which is already established for a certain question.

Why it is called “Null”? We call it Null Hypothesis because we always want to nullify the existing theory with something else. This is why we perform any test right? If we do not want to prove something new then why should we put effort into doing any new test?

In statistics, Null Hypothesis is denoted as H0

Alternative Hypothesis

The alternative hypothesis (H1) is the opposite of the null hypothesis. The alternative hypothesis is the hypothesis that we want to or try to prove.

For example, previously, people believed that the sun is rotating around the Earth until Galileo, an Italian astronomer, proved through his observations that the Earth is actually revolving around the sun. This discovery challenged the widely accepted belief of geocentrism, which people used to believe for centuries.

Here previously believed theory (the Sun is rotating around the Earth) is the Null Hypothesis and the theory Galileo wanted to prove (Earth is rotating around the Sun) is the Alternative Hypothesis. Here the test or observation Galileo perform can be called a Hypothesis testing.

In statistics, Alternative Hypothesis is denoted as H1

Hypothesis testing example

Let me now give your rela life example where you can implement hypothesis testing as a data science professional. Let’s say for a particular candy manufacturing company, it is believed that a candy machine makes chocolate bars that are an average of 5 grams. That candy manufacturing plant has been making chocolate bars for 10 years now suddenly a worker claims that the machine no longer makes 5 grams of each chocolate bar.

real-life-application-of-hypothesis-testing

So is the worker wrong or the machine is not functioning properly? To answer those questions the owner of that candy company should perform Hypothesis testing.

Here the Null Hypothesis is the statement “The machine is functioning correctly and making each chocolate bar 5 grams” and the alternative hypothesis is the statement of the worker “The machine no longer makes 5 grams of each chocolate bar”.

If we want to write above English text into math then we can write like below:

hypothesis-testing-in-statistics

Pass outcome of Hypothesis testing

There are only two outcomes of hypothesis testing in statistics those are:

  • Reject Null Hypothesis (H0) – the machine is producing each chocolate bar with 5g of weight
  • Fail to reject null hypothesis – In other words, alternative hypothesis (H1) is proven or new theory is established

Application of Hypothesis Testing

Like the above example, there are so many places we can implement hypothesis testing in real life problems. Let’s also see some other important areas where we can apply hypothesis testing.

  1. Medicine: Hypothesis testing used to determine effectiveness of new treatments or drugs.
  2. Marketing: You can understand the effectiveness of advertising campaigns or new product launches.
  3. Quality Control: Find out whether a manufacturing process is producing products within acceptable quality limits. This is the example I explained above.
  4. Environmental Science: You can test the effects of pollutants on ecosystems using hypothesis testing
Also Read:  Neural network from scratch in Python

Terminologies to know

Before calculating hypothesis testing you should know some statistical terminologies, let’s understand those.

Statistical Significance

Statistically Significant is a term used to describe whether a result from a statistical test or analysis is likely to be real or just by chance. It is the point where we can draw the line to make a decision. Again the decision is whether we are rejecting the null hypothesis or not.

Level of Confidence

The Level of Confidence (c) says that, how much confident are we when taking our decision to reject null hypothesis or fail to reject the null hypothesis. Level of confidence is measured as a probability value. Now when you are trying to prove something new (rejecting null hypo), you must have high confidence in your test so that people or statistics can believe in your new theory. Good confidence level is more than 95%.

Level of Significance

Level of Significance is a probability value used in hypothesis testing to determine if the results of a statistical test are statistically significant or not. This level of significance is also known as the p-value, which I will explain in a later post. The equation of LOS is as follows:

Level of Significance = 1 – Level of confidence

Now if for some test, level of confidence = 95%

Then, Level of Significance = 1 – 95 = 0.05

Similarly, if LOC = 99% then LOS = 0.01.

You can see sum of Level of Significance and Level of confidence is equal to 1 (LOS + LOC = 1). The sum should always be 1.

Good level of significance score is less than 0.05 (which is greater than 95% of LOS).

Level of Confidence vs Level of Significance

In general, there is no difference between level of confidence and level of significance. They are both the same thing. Different problems or research papers may use different parameters to indicate significance. But both will tell you the same thing, how sure are you to make any statistical decision?

Degrees of Freedom

In statistics, degrees of freedom (df) is the number of values in a sample that are free to vary after imposing certain restrictions or constraints.

In simple words, Degrees of freedom represents the number of independent observations or data points that can be used to estimate statistical parameters or test hypotheses. The more degrees of freedom available, the more reliable and accurate the statistical analysis is likely to be.

The equation for degrees of freedom is as follows:

DF = number of sample data (n) – 1

Hypothesis Testing Formula

Before we calculate hypothesis testing in hand from scratch, we should know the equation to test hypothesis. There are mainly two ways to calculate hypothesis those are t-test and z-test.

t-test equation

t-test equation is: t = ( x̅ – μ0 ) / (s / √n)

  • Where is sample mean
  • μ0 is already proven value (you can say null hypothesis value)
  • s sample standard deviation
  • n is the sample size

z-test equation

z-test equation is: Z = ( x̅ – μ0 ) / (σ / √n)

  • Where is sample mean
  • μ0 is already proven value (you can say null hypothesis value)
  • σ is the population standard deviation
  • n is the sample size

Difference between sample SD and population SD

For t-test, I mentioned sample standard deviation and for z-test, I mentioned population standard deviation. Now, what is the difference between sample SD and population SD?

The main difference between sample standard deviation (SD) and population standard deviation (SD) is the data set they are calculated from.

Sample standard deviation (SD) is calculated from a subset of the population data called a sample. A sample is a smaller group of data selected from the entire population for statistical analysis.

On the other hand Population standard deviation (SD) is calculated from the entire population or entire data set.

Z-Test vs. T-Test: When to Use Which?

The z-test and t-test are both statistical tests used to calculate hypothesis testing. The z-test is used when the population standard deviation is known and the sample size is large (typically n > 30).

On the other hand, t-test is used when the population standard deviation is unknown, and the sample size is small (typically n < 30).

Also Read:  Generate Synthetic Tabular Test Data in Python

In general, if the sample size is large and the population standard deviation is known then the z-test is preferred. If the sample size is small and the population standard deviation is unknown, then the t-test is more appropriate (you can use sample standard deviation).

How to calculate Hypothesis Testing in Excel

So we understand what is hypothesis. Now how to calculate it? or how do you actually determine if you reject the null hypothesis or not? For that, you need to perform some statistical test using some sample data.

Data Collection

Coming back to our main example of a candy-making factory where a worker suddenly made a statement that “the candy making machine is no longer makes 5 grams of chocolate bars”.

Now, this is a big and scary statement for the owner of that factory. So the owner assigned a quality inspector to inspect this.

The quality inspector checked the weight of ten chocolate randomly. I mentioned randomly which means the sample data should be normally distributed. Now the quality inspector can bring this randomness in various ways like he can check chocolate weight at different time of a day or week or check weight while different worker is working etc.

Now let’s say the quality inspector collected following ten random samples:

Candy NumberCandy Weight (in grams)
15.2
24.8
35
44.9
55.2
64.7
75
84.8
94.9
105.1

Selecting Test Type

As I explained earlier since the data or samle size is less and we do not know the population standard deviation so we will use T-test to calculate hypothesis for this problem.

Note: To solve almost all real life problem, we use t-test to calculate hypothesis testing because of lack of overall information (unknown population standard deviation).

Calculate required Parameters

So to apply t-test, now we need to calculate sample standard distribution for our example. You can do it easily in Excel. Now in excel, there are mainly two formulas for standard deviation. Those are:

  • STDEV.S => To calculate standard deviation for sample
  • STDEV.P +> To calculate population standard deviation

We can find our sample standard deviation value is 0.171 (using STDEV.S method in excel)

An our sample mean (x̅) = (5.2 + 4.8 + 5 + 4.9 + 5.2 + 4.7 + 5 + 4.8 + 4.9 + 5.1) / 10 = 4.96

So let’s now list down all required information to calculate hypothesis testing for our problem (measure likelihood of the worker’s claim). Informations are:

  • Already proven value (μ0) = 5 grams
  • Sample standard deviation (s) = 0.171
  • Number of samples collected (n) = 10
  • Degrees of Freedom (DF) = (n-1) = 9
  • Sample mean (x̅) = 4.96

Assuming that the company’s claim of a chocolate bar weight being 5 grams each is true, We need to prove that:

  • Either H0: μ=μ0 (null hypothesis is true – fail to reject null hypothesis – worker’s claimed information is not correct)
  • Or H1: μ≠μ0 (Rejecting null hypothesis – worker’s claim is correct)

Draw distribution graph

So our sample mean at 4.96 and our sample standard deviation at 0.171. Now if you draw the distribution of the possible values of our sample mean given the population mean is 5 (already proven weight of a chocolate bar which is 5 grams each). Assuming the null hypothesis is true, below would be the distribution:

standard-normal-distribution-gaussian-bell-graph-curve-concept-hypothesis-testing-gauss-distribution

So basically this hypothesis test is really going to be asking how extreme is this value of 4.96 (sample mean x̅) if the null hypothesis is true. Is it too extreme for us to accept null hypothesis or closer to population mean (already proven value 5) to reject null hypothesis? To answer this question we need to calculate a test statistic here using all provided or known informations.

Apply T test

Applying the T-test formula:

t = ( x̅ – μ0 ) / (s / √n)

t = (4.96 – 5) / (0.171 / √ 10) = -0.04/0.054 = -0.74074

Decision Rule

So we have our T score which is -0.74074. Now the next thing we need to do is to consider our decision rule. Like normal distribution, T distribution is also a bell curve type shape. The only difference is that the T distribution is defined by its number of degrees of freedom. In our example, degrees of freedom (DF) is 9.

So if degree of freedom is 9 and if we want 95% level of confidence or 0.05 level of significance, we can reject null hypothesis if our calculated T score lies in the green region on the negative side.

t-distribution-bell-curve-for-hypothesis-testing-with-005-level-of-significance

Output of Hypothesis Testing

Now from the T Critical values table, we can find that if degrees of freedom = 9 and Level of significance = 0.05, then T score should be greater than -1.833 (we will keep this value negative because our T score is negative) then only we can reject the null hypothesis.

Also Read:  Install MiniConda & setup Environment for Machine Learning

Note: This is just a one-tailed test as we are only looking at one direction from our expected value (in the negative direction). If we were looking for both the direction of distribution bell curve then we could call it a two-tailed test.

Below is the T Critical values table for your reference.

T distribution table

Conf. Level50%80%90%95%98%99%
One Tail0.250.10.050.0250.010.005
Two Tail0.50.20.10.050.020.01
df = 113.0786.31412.70631.82163.657
20.8161.8862.924.3036.9659.925
30.7651.6382.3533.1824.5415.841
40.7411.5332.1322.7763.7474.604
50.7271.4762.0152.5713.3654.032
60.7181.441.9432.4473.1433.707
70.7111.4151.8952.3652.9983.499
80.7061.3971.862.3062.8963.355
90.7031.3831.8332.2622.8213.25
100.71.3721.8122.2282.7643.169
110.6971.3631.7962.2012.7183.106
120.6951.3561.7822.1792.6813.055
130.6941.351.7712.162.653.012
140.6921.3451.7612.1452.6242.977
150.6911.3411.7532.1312.6022.947
160.691.3371.7462.122.5832.921
170.6891.3331.742.112.5672.898
180.6881.331.7342.1012.5522.878
190.6881.3281.7292.0932.5392.861
200.6871.3251.7252.0862.5282.845
210.6861.3231.7212.082.5182.831
220.6861.3211.7172.0742.5082.819
230.6851.3191.7142.0692.52.807
240.6851.3181.7112.0642.4922.797
250.6841.3161.7082.062.4852.787
260.6841.3151.7062.0562.4792.779
270.6841.3141.7032.0522.4732.771
280.6831.3131.7012.0482.4672.763
290.6831.3111.6992.0452.4622.756
300.6831.311.6972.0422.4572.75
400.6811.3031.6842.0212.4232.704
500.6791.2991.6762.0092.4032.678
600.6791.2961.67122.392.66
700.6781.2941.6671.9942.3812.648
800.6781.2921.6641.992.3742.639
900.6771.2911.6621.9872.3682.632
1000.6771.291.661.9842.3642.626
z0.6741.2821.6451.962.3262.576

The calculated T score for our example is -0.74074 > -1.833, therefore the null hypothesis is true (we failed to reject the null hypothesis). That means the claim by the worker is not true. The candy machine is working properly and it is still producing each chocolate bar around 5 grams.

Graph interpretation

We can see the above interpretation in the below T distribution bell curve:

output-decision-of-hypothesis-testing-calculation-in-hand-and-excel-from-scratch

End Note

In statistics, we should always test a hypothesis to make a proper judgment. Now there are two types of hypotheses. The first one is the Null Hypothesis, which is an already established concept. The second one is the Alternative Hypothesis, which is a new concept that we are trying to establish by proving the Null Hypothesis wrong.

We are making these decisions (rejecting null hypothesis or not) based on Level of confidence or Level of Significance score. Good score for those parameters are more than 95% for LOC or less than 0.05 for LOS.

In this post, I tried to explain Hypothesis testing with easy and real life examples. Please let me know if you have any questions or suggestions regarding this tutorial.

Leave a comment