Statistical Inference is the process of using data analysis to deduce properties of an underlying distribution of probability. Essentially, it's about making conclusions about a population based on a sample.
In statistical hypothesis testing, we use the Null Hypothesis ($H_{0}$) and the P-Value together to determine if a result is a real effect or just a lucky coincidence.
The null hypothesis is NOT a proven fact. It is just our starting stance of skepticism. Think of it like "Innocent until proven guilty" in a courtroom.
We assume the defendant is innocent ($H_0$: No Crime/Effect) not because we have proven they are innocent, but because we need a neutral starting point before looking at the evidence.
Example: If you are testing a new drug, we start by assuming "This drug does nothing" ($H_0$). We hold onto this assumption until the data (evidence) becomes so strong that we are forced to abandon it.
The p-value is a probability (ranging from 0 to 1) that represents how likely it is to observe your specific results if the null hypothesis were actually true. In simple words it is the probability that the null hypothesis is true.
We compare the p-value to a pre-set threshold called the Significance Level ($\alpha$), which is most commonly 0.05. If our P-Value is smaller than $\alpha$, we conclude the effect is real!
The curve represents the Null Hypothesis (e.g., the drug does nothing). Drag the slider to see how likely your experimental result would be if the drug actually had 0 effect.
Drag the slider to the right (Effect Strength > 2.0).
The Red Area is the P-Value. It shows how much of the "Zero Effect" distribution matches your result.
If the P-Value is tiny (Red area disappears), it means your result is too big to be a coincidence!
There isn't one single formula, but the general logic for finding P is:
Why Low P = Reject Null?
Think of P as the "Probability of Coincidence".
• If P = 0.03 (3%), it means there is only a 3% chance this happened by luck. That's too rare. So we assume it wasn't luck (Reject Null).
• If P = 0.40 (40%), it means there's a 40% chance this was just noise. That happens all the time. So we assume it was just noise (Keep Null).
Instead of a single number estimate, a Confidence Interval gives us a range that is likely to contain the true population parameter.
"95% Confidence" means: If we repeated this experiment 100 times, 95 of the calculated intervals would capture the true value.
The vertical white line is the True Population Mean (Hidden from the observer).
Click "Sample" to run an experiment and generate an interval.
While the P-Value is the universal "score," we calculate it using different tests depending on the type of data we have.
Use when: Comparing the averages (means) of two groups.
Use when: Comparing the averages of three or more groups.
Use when: Comparing categories / counts (not averages).