Understanding p-values using an example Unicorn Meta Zoo #1: Why another podcast? Announcing the arrival of Valued Associate #679: Cesar ManaraHow to interpret the results of a t test?Meaning of “was produced by the null hypothesis”Comparing and contrasting, p-values, significance levels and type I errorInterpreting p-values in Fisher vs Neyman-Pearson frameworksIs the p-postulate (equal p-values provide equal evidence against the null) true?Why are lower p-values not more evidence against the null? Arguments from Johansson 2011Are smaller p-values more convincing?Doubling &/or halving p-values for one- vs. two-tailed testsWhy are Type-II errors in hypothesis testing considered to be errors?Conceptual understanding of the P-valueWhy reject Null Hypothesis when p value< alpha?Where is my potential flaw in my Z-test of Proportions?
How do I check if a string is entirely made of the same substring?
What is the term for a person whose job is to place products on shelves in stores?
Why do games have consumables?
"My boss was furious with me and I have been fired" vs. "My boss was furious with me and I was fired"
Expansion//Explosion and Siren Stormtamer
Why does the Cisco show run command not show the full version, while the show version command does?
A strange hotel
What to do with someone that cheated their way through university and a PhD program?
"Whatever a Russian does, they end up making the Kalashnikov gun"? Are there any similar proverbs in English?
Implementing 3DES algorithm in Java: is my code secure?
Is it acceptable to use working hours to read general interest books?
Can I criticise the more senior developers around me for not writing clean code?
"Rubric" as meaning "signature" or "personal mark" -- is this accepted usage?
Is Electric Central Heating worth it if using Solar Panels?
Flattening the sub-lists
Reattaching fallen shelf to wall?
Rolling Stones Sway guitar solo chord function
What is it called when you ride around on your front wheel?
How to open locks without disable device?
A Dictionary or Encyclopedia of Fantasy or Fairy Tales from the 1960s
Israeli soda type drink
Holes in ElementMesh with ToElementMesh of ImplicitRegion
Seek and ye shall find
Multiple options vs single option UI
Understanding p-values using an example
Unicorn Meta Zoo #1: Why another podcast?
Announcing the arrival of Valued Associate #679: Cesar ManaraHow to interpret the results of a t test?Meaning of “was produced by the null hypothesis”Comparing and contrasting, p-values, significance levels and type I errorInterpreting p-values in Fisher vs Neyman-Pearson frameworksIs the p-postulate (equal p-values provide equal evidence against the null) true?Why are lower p-values not more evidence against the null? Arguments from Johansson 2011Are smaller p-values more convincing?Doubling &/or halving p-values for one- vs. two-tailed testsWhy are Type-II errors in hypothesis testing considered to be errors?Conceptual understanding of the P-valueWhy reject Null Hypothesis when p value< alpha?Where is my potential flaw in my Z-test of Proportions?
.everyoneloves__top-leaderboard:empty,.everyoneloves__mid-leaderboard:empty,.everyoneloves__bot-mid-leaderboard:empty margin-bottom:0;
$begingroup$
Definition of p-values: A p-value is a probability that provides a measure of the evidence against the null hypothesis provided by the sample. Smaller p-values indicate more evidence against null hypothesis. Can someone please explain this in simpler terms or in a language easy to understand?
I know there might already be tons of questions around understanding the interpretation of p-values, however I would ask the question in a very limited form and with the use of a specific example:
A manufacturing company fills up can with mean weight of 3 pounds, the level of significance is assumed to be 0.01
H(0) : u >= 3 -- Null hypotheses
H(a) : u < 3 -- Alternate hypotheses
We are trying to perform a one tailed test for the case where the population standard deviation is known, so for a sample mean of 2.92 and a standard error of 0.03, we get the z-score as -2.67, giving us the probability (p-value) of 0.0038 or 0.38% that the sample mean would be equal to or lower than 2.92.
Since the probability of getting a sample mean equal or less than 2.92 is 0.38%, which is very small, doesn't it mean that we should accept the null hypotheses? As the chances of getting a mean of 2.92 from a sample is only 0.38%.
Or am I completely missing something here?
Edit - It has been three days now since I tried understanding hypothesis testing and I think I am almost there, I will try to articulate what I have understood so far and then let me know if there are still any gaps in my understanding
p-values measure the likelihood of obtaining the sample mean that we obtained given that the null hypothesis is true. So for the example that I mentioned, the probability of obtaining a sample mean of 2.92 is 0.038 if that population's mean is 3 (as assumed by the null hypothesis).
Now there could be two reasons for obtaining means of 2.92:
- The assumed population mean (i.e., the null hypothesis) is not correct, or
- the population mean is 3 but due to a sampling error / an unlikely sample we got a mean of 2.92.
Now if we select statement 1, we run the chance of making type 1 error and this is where the level of significance comes into play. Using the level of significance we can see if we can reject the null hypothesis or cannot reject null hypothesis.
hypothesis-testing statistical-significance p-value
New contributor
$endgroup$
add a comment |
$begingroup$
Definition of p-values: A p-value is a probability that provides a measure of the evidence against the null hypothesis provided by the sample. Smaller p-values indicate more evidence against null hypothesis. Can someone please explain this in simpler terms or in a language easy to understand?
I know there might already be tons of questions around understanding the interpretation of p-values, however I would ask the question in a very limited form and with the use of a specific example:
A manufacturing company fills up can with mean weight of 3 pounds, the level of significance is assumed to be 0.01
H(0) : u >= 3 -- Null hypotheses
H(a) : u < 3 -- Alternate hypotheses
We are trying to perform a one tailed test for the case where the population standard deviation is known, so for a sample mean of 2.92 and a standard error of 0.03, we get the z-score as -2.67, giving us the probability (p-value) of 0.0038 or 0.38% that the sample mean would be equal to or lower than 2.92.
Since the probability of getting a sample mean equal or less than 2.92 is 0.38%, which is very small, doesn't it mean that we should accept the null hypotheses? As the chances of getting a mean of 2.92 from a sample is only 0.38%.
Or am I completely missing something here?
Edit - It has been three days now since I tried understanding hypothesis testing and I think I am almost there, I will try to articulate what I have understood so far and then let me know if there are still any gaps in my understanding
p-values measure the likelihood of obtaining the sample mean that we obtained given that the null hypothesis is true. So for the example that I mentioned, the probability of obtaining a sample mean of 2.92 is 0.038 if that population's mean is 3 (as assumed by the null hypothesis).
Now there could be two reasons for obtaining means of 2.92:
- The assumed population mean (i.e., the null hypothesis) is not correct, or
- the population mean is 3 but due to a sampling error / an unlikely sample we got a mean of 2.92.
Now if we select statement 1, we run the chance of making type 1 error and this is where the level of significance comes into play. Using the level of significance we can see if we can reject the null hypothesis or cannot reject null hypothesis.
hypothesis-testing statistical-significance p-value
New contributor
$endgroup$
$begingroup$
You say the population standard deviation $sigma$ is known. Can you provide the known value? The terminology "correction factor" is not familiar to me; can you give a formula for finding that? // The sample mean $bar X = 2.92$ is below the hypothetical population mean $mu_0 = 3.$ The issue is whether it is enough smaller to warrant rejecting the null hypothesis.
$endgroup$
– BruceET
Apr 20 at 13:20
$begingroup$
The population standard deviation is .18 and sample size is 36, hence the correction factor is 0.18/sqrt(36) equals 0.03
$endgroup$
– Rohit Saluja
Apr 20 at 13:24
2
$begingroup$
Thanks for the additional information. The usual terminology is to call $sigma/sqrtn$ the 'standard error'.
$endgroup$
– BruceET
Apr 20 at 13:27
$begingroup$
@BruceET - the issue is if it is enough smaller to reject the null hypotheses, however the probability of sample mean of being less than or equal to 2.92 is only 0.0038, so can't we say that the probability of sample mean of less than 3 is very less hence we support null hypotheses.
$endgroup$
– Rohit Saluja
Apr 20 at 13:39
add a comment |
$begingroup$
Definition of p-values: A p-value is a probability that provides a measure of the evidence against the null hypothesis provided by the sample. Smaller p-values indicate more evidence against null hypothesis. Can someone please explain this in simpler terms or in a language easy to understand?
I know there might already be tons of questions around understanding the interpretation of p-values, however I would ask the question in a very limited form and with the use of a specific example:
A manufacturing company fills up can with mean weight of 3 pounds, the level of significance is assumed to be 0.01
H(0) : u >= 3 -- Null hypotheses
H(a) : u < 3 -- Alternate hypotheses
We are trying to perform a one tailed test for the case where the population standard deviation is known, so for a sample mean of 2.92 and a standard error of 0.03, we get the z-score as -2.67, giving us the probability (p-value) of 0.0038 or 0.38% that the sample mean would be equal to or lower than 2.92.
Since the probability of getting a sample mean equal or less than 2.92 is 0.38%, which is very small, doesn't it mean that we should accept the null hypotheses? As the chances of getting a mean of 2.92 from a sample is only 0.38%.
Or am I completely missing something here?
Edit - It has been three days now since I tried understanding hypothesis testing and I think I am almost there, I will try to articulate what I have understood so far and then let me know if there are still any gaps in my understanding
p-values measure the likelihood of obtaining the sample mean that we obtained given that the null hypothesis is true. So for the example that I mentioned, the probability of obtaining a sample mean of 2.92 is 0.038 if that population's mean is 3 (as assumed by the null hypothesis).
Now there could be two reasons for obtaining means of 2.92:
- The assumed population mean (i.e., the null hypothesis) is not correct, or
- the population mean is 3 but due to a sampling error / an unlikely sample we got a mean of 2.92.
Now if we select statement 1, we run the chance of making type 1 error and this is where the level of significance comes into play. Using the level of significance we can see if we can reject the null hypothesis or cannot reject null hypothesis.
hypothesis-testing statistical-significance p-value
New contributor
$endgroup$
Definition of p-values: A p-value is a probability that provides a measure of the evidence against the null hypothesis provided by the sample. Smaller p-values indicate more evidence against null hypothesis. Can someone please explain this in simpler terms or in a language easy to understand?
I know there might already be tons of questions around understanding the interpretation of p-values, however I would ask the question in a very limited form and with the use of a specific example:
A manufacturing company fills up can with mean weight of 3 pounds, the level of significance is assumed to be 0.01
H(0) : u >= 3 -- Null hypotheses
H(a) : u < 3 -- Alternate hypotheses
We are trying to perform a one tailed test for the case where the population standard deviation is known, so for a sample mean of 2.92 and a standard error of 0.03, we get the z-score as -2.67, giving us the probability (p-value) of 0.0038 or 0.38% that the sample mean would be equal to or lower than 2.92.
Since the probability of getting a sample mean equal or less than 2.92 is 0.38%, which is very small, doesn't it mean that we should accept the null hypotheses? As the chances of getting a mean of 2.92 from a sample is only 0.38%.
Or am I completely missing something here?
Edit - It has been three days now since I tried understanding hypothesis testing and I think I am almost there, I will try to articulate what I have understood so far and then let me know if there are still any gaps in my understanding
p-values measure the likelihood of obtaining the sample mean that we obtained given that the null hypothesis is true. So for the example that I mentioned, the probability of obtaining a sample mean of 2.92 is 0.038 if that population's mean is 3 (as assumed by the null hypothesis).
Now there could be two reasons for obtaining means of 2.92:
- The assumed population mean (i.e., the null hypothesis) is not correct, or
- the population mean is 3 but due to a sampling error / an unlikely sample we got a mean of 2.92.
Now if we select statement 1, we run the chance of making type 1 error and this is where the level of significance comes into play. Using the level of significance we can see if we can reject the null hypothesis or cannot reject null hypothesis.
hypothesis-testing statistical-significance p-value
hypothesis-testing statistical-significance p-value
New contributor
New contributor
edited 2 days ago
gung♦
109k34268539
109k34268539
New contributor
asked Apr 20 at 12:34
Rohit SalujaRohit Saluja
1284
1284
New contributor
New contributor
$begingroup$
You say the population standard deviation $sigma$ is known. Can you provide the known value? The terminology "correction factor" is not familiar to me; can you give a formula for finding that? // The sample mean $bar X = 2.92$ is below the hypothetical population mean $mu_0 = 3.$ The issue is whether it is enough smaller to warrant rejecting the null hypothesis.
$endgroup$
– BruceET
Apr 20 at 13:20
$begingroup$
The population standard deviation is .18 and sample size is 36, hence the correction factor is 0.18/sqrt(36) equals 0.03
$endgroup$
– Rohit Saluja
Apr 20 at 13:24
2
$begingroup$
Thanks for the additional information. The usual terminology is to call $sigma/sqrtn$ the 'standard error'.
$endgroup$
– BruceET
Apr 20 at 13:27
$begingroup$
@BruceET - the issue is if it is enough smaller to reject the null hypotheses, however the probability of sample mean of being less than or equal to 2.92 is only 0.0038, so can't we say that the probability of sample mean of less than 3 is very less hence we support null hypotheses.
$endgroup$
– Rohit Saluja
Apr 20 at 13:39
add a comment |
$begingroup$
You say the population standard deviation $sigma$ is known. Can you provide the known value? The terminology "correction factor" is not familiar to me; can you give a formula for finding that? // The sample mean $bar X = 2.92$ is below the hypothetical population mean $mu_0 = 3.$ The issue is whether it is enough smaller to warrant rejecting the null hypothesis.
$endgroup$
– BruceET
Apr 20 at 13:20
$begingroup$
The population standard deviation is .18 and sample size is 36, hence the correction factor is 0.18/sqrt(36) equals 0.03
$endgroup$
– Rohit Saluja
Apr 20 at 13:24
2
$begingroup$
Thanks for the additional information. The usual terminology is to call $sigma/sqrtn$ the 'standard error'.
$endgroup$
– BruceET
Apr 20 at 13:27
$begingroup$
@BruceET - the issue is if it is enough smaller to reject the null hypotheses, however the probability of sample mean of being less than or equal to 2.92 is only 0.0038, so can't we say that the probability of sample mean of less than 3 is very less hence we support null hypotheses.
$endgroup$
– Rohit Saluja
Apr 20 at 13:39
$begingroup$
You say the population standard deviation $sigma$ is known. Can you provide the known value? The terminology "correction factor" is not familiar to me; can you give a formula for finding that? // The sample mean $bar X = 2.92$ is below the hypothetical population mean $mu_0 = 3.$ The issue is whether it is enough smaller to warrant rejecting the null hypothesis.
$endgroup$
– BruceET
Apr 20 at 13:20
$begingroup$
You say the population standard deviation $sigma$ is known. Can you provide the known value? The terminology "correction factor" is not familiar to me; can you give a formula for finding that? // The sample mean $bar X = 2.92$ is below the hypothetical population mean $mu_0 = 3.$ The issue is whether it is enough smaller to warrant rejecting the null hypothesis.
$endgroup$
– BruceET
Apr 20 at 13:20
$begingroup$
The population standard deviation is .18 and sample size is 36, hence the correction factor is 0.18/sqrt(36) equals 0.03
$endgroup$
– Rohit Saluja
Apr 20 at 13:24
$begingroup$
The population standard deviation is .18 and sample size is 36, hence the correction factor is 0.18/sqrt(36) equals 0.03
$endgroup$
– Rohit Saluja
Apr 20 at 13:24
2
2
$begingroup$
Thanks for the additional information. The usual terminology is to call $sigma/sqrtn$ the 'standard error'.
$endgroup$
– BruceET
Apr 20 at 13:27
$begingroup$
Thanks for the additional information. The usual terminology is to call $sigma/sqrtn$ the 'standard error'.
$endgroup$
– BruceET
Apr 20 at 13:27
$begingroup$
@BruceET - the issue is if it is enough smaller to reject the null hypotheses, however the probability of sample mean of being less than or equal to 2.92 is only 0.0038, so can't we say that the probability of sample mean of less than 3 is very less hence we support null hypotheses.
$endgroup$
– Rohit Saluja
Apr 20 at 13:39
$begingroup$
@BruceET - the issue is if it is enough smaller to reject the null hypotheses, however the probability of sample mean of being less than or equal to 2.92 is only 0.0038, so can't we say that the probability of sample mean of less than 3 is very less hence we support null hypotheses.
$endgroup$
– Rohit Saluja
Apr 20 at 13:39
add a comment |
4 Answers
4
active
oldest
votes
$begingroup$
Imagine you could measure the weight of all cans that the manufacturing company has ever made and the mean would be $2.87$ pounds. Then imagine you would take 10 cans randomly and see how much they weight. It is unlikely to get the exact mean of all cans ($2.87$ pounds), hence you end up with a mean of $2.65$, for example. If you would do that again and again - taking 10 cans and measuring the weight - you would get a distribution of means. The best guess about the true mean is the mean of the distribution you obtained. Extreme values like $1.9$ or $3.5$ pounds will be unlikely and even more extreme values will be even more unlikely.
Doing significance tests usually means that you look how likely the mean you observed is if you assume that your sample was drawn from a population with mean zero. If the mean that you observed is very unlikely you would decide to discard the null hypothesis. The only difference between what I have said so far and your example is that you assume the null hypothesis a mean of $ge 3$. So the $0.38%$ you report say that the probability of getting your mean of $2.92$ from a population with a mean of $ge 3$ is so unlikely that you would discard the null hypothesis and accept the alternative hypothesis which is $<3$. Your evidence indicate that the cans weight less than $3$ pounds.
This means it is the opposite: having a $p$ of $0.38%$ as you report doesn't mean you have to keep the null hypothesis because your result is so unlikely but it means that you can discard the null hypothesis because your data was very unlikely a randomly drawn sample from a population with a mean of $3$ (i.e., your data would be very unlikely given that the null hypothesis is true).
New contributor
$endgroup$
$begingroup$
Comments are not for extended discussion; this conversation has been moved to chat.
$endgroup$
– gung♦
yesterday
add a comment |
$begingroup$
Here is a figure that shows your problem on two scales:
at left is the original scale in terms of pounds; at right
is the standard of z-scale often used in testing.
To begin, let's look at your problem in terms of the fixed significance level $alpha = 0.01 = 1%.$ In the right-hand
panel, your $Z$-score is shown at the heavy vertical bar at
$-2.67.$ The "critical value" for a test at the 1% level is
shown by the vertical dotted line at $-2.326,$ which cuts
1% of the probability from the lower tail of the standard normal
distribution.
Because the $Z$-score is to the left of the
critical value, one rejects the null hypothesis at level $alpha = 1%.$ The P-value is the probability under the standard normal curve to the left of the heavy blue line. That area is smaller than $1%,$ so in terms of P-values, we reject $H_0$ when the P-value is smaller than $1%.$
You can see that the left-hand plot is the same as the right-hand plot, except for scale. It is not possible to make a printed normal table for all possible normal distributions.
By converting to $Z$-scores we can always use a single printed table for the 'standard' normal distribution, which has mean 0 and standard deviation 1.
If we were going to do this production-monitoring procedure repeatedly with $n = 36$ observations each time, then we could find the critical value on the 'pound' scale; it is
at 2.581 pounds. (That's because $(2.581 - 3)/.18 = -2.236,$ where the
$0.18$ is the standard error.) Then we could turn the testing
job over to a non-statistician, with instructions: "If the average weight for 36 cans is less than 2.581 pounds, let me know because we aren't putting enough stuff in our cans." (Or if we can't even trust the non-statistician with averages, the criterion might be a total weight less than 92.92 pounds.)
$endgroup$
add a comment |
$begingroup$
Since your question is actually quite precise, I would like to keep it rather concise.
Definition of p-value: the p-value is the probability of the data (or even more extrem data) given the null hypothesis is actually true.
If this probability is high, then there is no reason why we should reject the null hypothesis: the data is perfectly in line with the null hypothesis. If the p-value is small, then the data seems implausible given the null hypothesis. The more implausible the data, the stronger our evidence against the null.
A level of significance of 0.01 means: to reject the null hypothesis, the probability of the data must be less than 1%. If the null hypothesis is actually true, we have therefore 1% chance to see data, which is so implausible that we would wrongly reject the null hypothesis.
Regarding your example: there is only 0.38% chance to see this data, if the null hypothesis is true, which is below our threshold of significance. Hence, the data seems very unlikely, and therefore we conclude that we don't believe in the null hypothesis anymore.
$endgroup$
add a comment |
$begingroup$
Assume the significance level is $alpha$, which when talking about the null hypothesis, we are usually looking at 5% or 1% and so on.
In simple terms: p-value is the smallest $alpha$ at which we reject the null hypothesis.
So, when your p-value is 0.15, then we accept the null hypothesis when $alpha$ is 5% (or our confidence interval is 90%). But change that to only have a confidence interval of 60% and you reject your null hypothesis. Similarly, when your p-value=0.0038, it means you accept the null hypothesis under any value smaller than < 0.38%. That's why you compare p-value with $alpha$ and if p-value < $alpha$, you say that you cannot accept the null hypothesis.
New contributor
$endgroup$
add a comment |
Your Answer
StackExchange.ready(function()
var channelOptions =
tags: "".split(" "),
id: "65"
;
initTagRenderer("".split(" "), "".split(" "), channelOptions);
StackExchange.using("externalEditor", function()
// Have to fire editor after snippets, if snippets enabled
if (StackExchange.settings.snippets.snippetsEnabled)
StackExchange.using("snippets", function()
createEditor();
);
else
createEditor();
);
function createEditor()
StackExchange.prepareEditor(
heartbeatType: 'answer',
autoActivateHeartbeat: false,
convertImagesToLinks: false,
noModals: true,
showLowRepImageUploadWarning: true,
reputationToPostImages: null,
bindNavPrevention: true,
postfix: "",
imageUploader:
brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
allowUrls: true
,
onDemand: true,
discardSelector: ".discard-answer"
,immediatelyShowMarkdownHelp:true
);
);
Rohit Saluja is a new contributor. Be nice, and check out our Code of Conduct.
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function ()
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstats.stackexchange.com%2fquestions%2f404119%2funderstanding-p-values-using-an-example%23new-answer', 'question_page');
);
Post as a guest
Required, but never shown
4 Answers
4
active
oldest
votes
4 Answers
4
active
oldest
votes
active
oldest
votes
active
oldest
votes
$begingroup$
Imagine you could measure the weight of all cans that the manufacturing company has ever made and the mean would be $2.87$ pounds. Then imagine you would take 10 cans randomly and see how much they weight. It is unlikely to get the exact mean of all cans ($2.87$ pounds), hence you end up with a mean of $2.65$, for example. If you would do that again and again - taking 10 cans and measuring the weight - you would get a distribution of means. The best guess about the true mean is the mean of the distribution you obtained. Extreme values like $1.9$ or $3.5$ pounds will be unlikely and even more extreme values will be even more unlikely.
Doing significance tests usually means that you look how likely the mean you observed is if you assume that your sample was drawn from a population with mean zero. If the mean that you observed is very unlikely you would decide to discard the null hypothesis. The only difference between what I have said so far and your example is that you assume the null hypothesis a mean of $ge 3$. So the $0.38%$ you report say that the probability of getting your mean of $2.92$ from a population with a mean of $ge 3$ is so unlikely that you would discard the null hypothesis and accept the alternative hypothesis which is $<3$. Your evidence indicate that the cans weight less than $3$ pounds.
This means it is the opposite: having a $p$ of $0.38%$ as you report doesn't mean you have to keep the null hypothesis because your result is so unlikely but it means that you can discard the null hypothesis because your data was very unlikely a randomly drawn sample from a population with a mean of $3$ (i.e., your data would be very unlikely given that the null hypothesis is true).
New contributor
$endgroup$
$begingroup$
Comments are not for extended discussion; this conversation has been moved to chat.
$endgroup$
– gung♦
yesterday
add a comment |
$begingroup$
Imagine you could measure the weight of all cans that the manufacturing company has ever made and the mean would be $2.87$ pounds. Then imagine you would take 10 cans randomly and see how much they weight. It is unlikely to get the exact mean of all cans ($2.87$ pounds), hence you end up with a mean of $2.65$, for example. If you would do that again and again - taking 10 cans and measuring the weight - you would get a distribution of means. The best guess about the true mean is the mean of the distribution you obtained. Extreme values like $1.9$ or $3.5$ pounds will be unlikely and even more extreme values will be even more unlikely.
Doing significance tests usually means that you look how likely the mean you observed is if you assume that your sample was drawn from a population with mean zero. If the mean that you observed is very unlikely you would decide to discard the null hypothesis. The only difference between what I have said so far and your example is that you assume the null hypothesis a mean of $ge 3$. So the $0.38%$ you report say that the probability of getting your mean of $2.92$ from a population with a mean of $ge 3$ is so unlikely that you would discard the null hypothesis and accept the alternative hypothesis which is $<3$. Your evidence indicate that the cans weight less than $3$ pounds.
This means it is the opposite: having a $p$ of $0.38%$ as you report doesn't mean you have to keep the null hypothesis because your result is so unlikely but it means that you can discard the null hypothesis because your data was very unlikely a randomly drawn sample from a population with a mean of $3$ (i.e., your data would be very unlikely given that the null hypothesis is true).
New contributor
$endgroup$
$begingroup$
Comments are not for extended discussion; this conversation has been moved to chat.
$endgroup$
– gung♦
yesterday
add a comment |
$begingroup$
Imagine you could measure the weight of all cans that the manufacturing company has ever made and the mean would be $2.87$ pounds. Then imagine you would take 10 cans randomly and see how much they weight. It is unlikely to get the exact mean of all cans ($2.87$ pounds), hence you end up with a mean of $2.65$, for example. If you would do that again and again - taking 10 cans and measuring the weight - you would get a distribution of means. The best guess about the true mean is the mean of the distribution you obtained. Extreme values like $1.9$ or $3.5$ pounds will be unlikely and even more extreme values will be even more unlikely.
Doing significance tests usually means that you look how likely the mean you observed is if you assume that your sample was drawn from a population with mean zero. If the mean that you observed is very unlikely you would decide to discard the null hypothesis. The only difference between what I have said so far and your example is that you assume the null hypothesis a mean of $ge 3$. So the $0.38%$ you report say that the probability of getting your mean of $2.92$ from a population with a mean of $ge 3$ is so unlikely that you would discard the null hypothesis and accept the alternative hypothesis which is $<3$. Your evidence indicate that the cans weight less than $3$ pounds.
This means it is the opposite: having a $p$ of $0.38%$ as you report doesn't mean you have to keep the null hypothesis because your result is so unlikely but it means that you can discard the null hypothesis because your data was very unlikely a randomly drawn sample from a population with a mean of $3$ (i.e., your data would be very unlikely given that the null hypothesis is true).
New contributor
$endgroup$
Imagine you could measure the weight of all cans that the manufacturing company has ever made and the mean would be $2.87$ pounds. Then imagine you would take 10 cans randomly and see how much they weight. It is unlikely to get the exact mean of all cans ($2.87$ pounds), hence you end up with a mean of $2.65$, for example. If you would do that again and again - taking 10 cans and measuring the weight - you would get a distribution of means. The best guess about the true mean is the mean of the distribution you obtained. Extreme values like $1.9$ or $3.5$ pounds will be unlikely and even more extreme values will be even more unlikely.
Doing significance tests usually means that you look how likely the mean you observed is if you assume that your sample was drawn from a population with mean zero. If the mean that you observed is very unlikely you would decide to discard the null hypothesis. The only difference between what I have said so far and your example is that you assume the null hypothesis a mean of $ge 3$. So the $0.38%$ you report say that the probability of getting your mean of $2.92$ from a population with a mean of $ge 3$ is so unlikely that you would discard the null hypothesis and accept the alternative hypothesis which is $<3$. Your evidence indicate that the cans weight less than $3$ pounds.
This means it is the opposite: having a $p$ of $0.38%$ as you report doesn't mean you have to keep the null hypothesis because your result is so unlikely but it means that you can discard the null hypothesis because your data was very unlikely a randomly drawn sample from a population with a mean of $3$ (i.e., your data would be very unlikely given that the null hypothesis is true).
New contributor
edited yesterday
gung♦
109k34268539
109k34268539
New contributor
answered Apr 20 at 13:44
stats.and.rstats.and.r
3478
3478
New contributor
New contributor
$begingroup$
Comments are not for extended discussion; this conversation has been moved to chat.
$endgroup$
– gung♦
yesterday
add a comment |
$begingroup$
Comments are not for extended discussion; this conversation has been moved to chat.
$endgroup$
– gung♦
yesterday
$begingroup$
Comments are not for extended discussion; this conversation has been moved to chat.
$endgroup$
– gung♦
yesterday
$begingroup$
Comments are not for extended discussion; this conversation has been moved to chat.
$endgroup$
– gung♦
yesterday
add a comment |
$begingroup$
Here is a figure that shows your problem on two scales:
at left is the original scale in terms of pounds; at right
is the standard of z-scale often used in testing.
To begin, let's look at your problem in terms of the fixed significance level $alpha = 0.01 = 1%.$ In the right-hand
panel, your $Z$-score is shown at the heavy vertical bar at
$-2.67.$ The "critical value" for a test at the 1% level is
shown by the vertical dotted line at $-2.326,$ which cuts
1% of the probability from the lower tail of the standard normal
distribution.
Because the $Z$-score is to the left of the
critical value, one rejects the null hypothesis at level $alpha = 1%.$ The P-value is the probability under the standard normal curve to the left of the heavy blue line. That area is smaller than $1%,$ so in terms of P-values, we reject $H_0$ when the P-value is smaller than $1%.$
You can see that the left-hand plot is the same as the right-hand plot, except for scale. It is not possible to make a printed normal table for all possible normal distributions.
By converting to $Z$-scores we can always use a single printed table for the 'standard' normal distribution, which has mean 0 and standard deviation 1.
If we were going to do this production-monitoring procedure repeatedly with $n = 36$ observations each time, then we could find the critical value on the 'pound' scale; it is
at 2.581 pounds. (That's because $(2.581 - 3)/.18 = -2.236,$ where the
$0.18$ is the standard error.) Then we could turn the testing
job over to a non-statistician, with instructions: "If the average weight for 36 cans is less than 2.581 pounds, let me know because we aren't putting enough stuff in our cans." (Or if we can't even trust the non-statistician with averages, the criterion might be a total weight less than 92.92 pounds.)
$endgroup$
add a comment |
$begingroup$
Here is a figure that shows your problem on two scales:
at left is the original scale in terms of pounds; at right
is the standard of z-scale often used in testing.
To begin, let's look at your problem in terms of the fixed significance level $alpha = 0.01 = 1%.$ In the right-hand
panel, your $Z$-score is shown at the heavy vertical bar at
$-2.67.$ The "critical value" for a test at the 1% level is
shown by the vertical dotted line at $-2.326,$ which cuts
1% of the probability from the lower tail of the standard normal
distribution.
Because the $Z$-score is to the left of the
critical value, one rejects the null hypothesis at level $alpha = 1%.$ The P-value is the probability under the standard normal curve to the left of the heavy blue line. That area is smaller than $1%,$ so in terms of P-values, we reject $H_0$ when the P-value is smaller than $1%.$
You can see that the left-hand plot is the same as the right-hand plot, except for scale. It is not possible to make a printed normal table for all possible normal distributions.
By converting to $Z$-scores we can always use a single printed table for the 'standard' normal distribution, which has mean 0 and standard deviation 1.
If we were going to do this production-monitoring procedure repeatedly with $n = 36$ observations each time, then we could find the critical value on the 'pound' scale; it is
at 2.581 pounds. (That's because $(2.581 - 3)/.18 = -2.236,$ where the
$0.18$ is the standard error.) Then we could turn the testing
job over to a non-statistician, with instructions: "If the average weight for 36 cans is less than 2.581 pounds, let me know because we aren't putting enough stuff in our cans." (Or if we can't even trust the non-statistician with averages, the criterion might be a total weight less than 92.92 pounds.)
$endgroup$
add a comment |
$begingroup$
Here is a figure that shows your problem on two scales:
at left is the original scale in terms of pounds; at right
is the standard of z-scale often used in testing.
To begin, let's look at your problem in terms of the fixed significance level $alpha = 0.01 = 1%.$ In the right-hand
panel, your $Z$-score is shown at the heavy vertical bar at
$-2.67.$ The "critical value" for a test at the 1% level is
shown by the vertical dotted line at $-2.326,$ which cuts
1% of the probability from the lower tail of the standard normal
distribution.
Because the $Z$-score is to the left of the
critical value, one rejects the null hypothesis at level $alpha = 1%.$ The P-value is the probability under the standard normal curve to the left of the heavy blue line. That area is smaller than $1%,$ so in terms of P-values, we reject $H_0$ when the P-value is smaller than $1%.$
You can see that the left-hand plot is the same as the right-hand plot, except for scale. It is not possible to make a printed normal table for all possible normal distributions.
By converting to $Z$-scores we can always use a single printed table for the 'standard' normal distribution, which has mean 0 and standard deviation 1.
If we were going to do this production-monitoring procedure repeatedly with $n = 36$ observations each time, then we could find the critical value on the 'pound' scale; it is
at 2.581 pounds. (That's because $(2.581 - 3)/.18 = -2.236,$ where the
$0.18$ is the standard error.) Then we could turn the testing
job over to a non-statistician, with instructions: "If the average weight for 36 cans is less than 2.581 pounds, let me know because we aren't putting enough stuff in our cans." (Or if we can't even trust the non-statistician with averages, the criterion might be a total weight less than 92.92 pounds.)
$endgroup$
Here is a figure that shows your problem on two scales:
at left is the original scale in terms of pounds; at right
is the standard of z-scale often used in testing.
To begin, let's look at your problem in terms of the fixed significance level $alpha = 0.01 = 1%.$ In the right-hand
panel, your $Z$-score is shown at the heavy vertical bar at
$-2.67.$ The "critical value" for a test at the 1% level is
shown by the vertical dotted line at $-2.326,$ which cuts
1% of the probability from the lower tail of the standard normal
distribution.
Because the $Z$-score is to the left of the
critical value, one rejects the null hypothesis at level $alpha = 1%.$ The P-value is the probability under the standard normal curve to the left of the heavy blue line. That area is smaller than $1%,$ so in terms of P-values, we reject $H_0$ when the P-value is smaller than $1%.$
You can see that the left-hand plot is the same as the right-hand plot, except for scale. It is not possible to make a printed normal table for all possible normal distributions.
By converting to $Z$-scores we can always use a single printed table for the 'standard' normal distribution, which has mean 0 and standard deviation 1.
If we were going to do this production-monitoring procedure repeatedly with $n = 36$ observations each time, then we could find the critical value on the 'pound' scale; it is
at 2.581 pounds. (That's because $(2.581 - 3)/.18 = -2.236,$ where the
$0.18$ is the standard error.) Then we could turn the testing
job over to a non-statistician, with instructions: "If the average weight for 36 cans is less than 2.581 pounds, let me know because we aren't putting enough stuff in our cans." (Or if we can't even trust the non-statistician with averages, the criterion might be a total weight less than 92.92 pounds.)
edited Apr 20 at 15:23
answered Apr 20 at 14:31
BruceETBruceET
7,2711721
7,2711721
add a comment |
add a comment |
$begingroup$
Since your question is actually quite precise, I would like to keep it rather concise.
Definition of p-value: the p-value is the probability of the data (or even more extrem data) given the null hypothesis is actually true.
If this probability is high, then there is no reason why we should reject the null hypothesis: the data is perfectly in line with the null hypothesis. If the p-value is small, then the data seems implausible given the null hypothesis. The more implausible the data, the stronger our evidence against the null.
A level of significance of 0.01 means: to reject the null hypothesis, the probability of the data must be less than 1%. If the null hypothesis is actually true, we have therefore 1% chance to see data, which is so implausible that we would wrongly reject the null hypothesis.
Regarding your example: there is only 0.38% chance to see this data, if the null hypothesis is true, which is below our threshold of significance. Hence, the data seems very unlikely, and therefore we conclude that we don't believe in the null hypothesis anymore.
$endgroup$
add a comment |
$begingroup$
Since your question is actually quite precise, I would like to keep it rather concise.
Definition of p-value: the p-value is the probability of the data (or even more extrem data) given the null hypothesis is actually true.
If this probability is high, then there is no reason why we should reject the null hypothesis: the data is perfectly in line with the null hypothesis. If the p-value is small, then the data seems implausible given the null hypothesis. The more implausible the data, the stronger our evidence against the null.
A level of significance of 0.01 means: to reject the null hypothesis, the probability of the data must be less than 1%. If the null hypothesis is actually true, we have therefore 1% chance to see data, which is so implausible that we would wrongly reject the null hypothesis.
Regarding your example: there is only 0.38% chance to see this data, if the null hypothesis is true, which is below our threshold of significance. Hence, the data seems very unlikely, and therefore we conclude that we don't believe in the null hypothesis anymore.
$endgroup$
add a comment |
$begingroup$
Since your question is actually quite precise, I would like to keep it rather concise.
Definition of p-value: the p-value is the probability of the data (or even more extrem data) given the null hypothesis is actually true.
If this probability is high, then there is no reason why we should reject the null hypothesis: the data is perfectly in line with the null hypothesis. If the p-value is small, then the data seems implausible given the null hypothesis. The more implausible the data, the stronger our evidence against the null.
A level of significance of 0.01 means: to reject the null hypothesis, the probability of the data must be less than 1%. If the null hypothesis is actually true, we have therefore 1% chance to see data, which is so implausible that we would wrongly reject the null hypothesis.
Regarding your example: there is only 0.38% chance to see this data, if the null hypothesis is true, which is below our threshold of significance. Hence, the data seems very unlikely, and therefore we conclude that we don't believe in the null hypothesis anymore.
$endgroup$
Since your question is actually quite precise, I would like to keep it rather concise.
Definition of p-value: the p-value is the probability of the data (or even more extrem data) given the null hypothesis is actually true.
If this probability is high, then there is no reason why we should reject the null hypothesis: the data is perfectly in line with the null hypothesis. If the p-value is small, then the data seems implausible given the null hypothesis. The more implausible the data, the stronger our evidence against the null.
A level of significance of 0.01 means: to reject the null hypothesis, the probability of the data must be less than 1%. If the null hypothesis is actually true, we have therefore 1% chance to see data, which is so implausible that we would wrongly reject the null hypothesis.
Regarding your example: there is only 0.38% chance to see this data, if the null hypothesis is true, which is below our threshold of significance. Hence, the data seems very unlikely, and therefore we conclude that we don't believe in the null hypothesis anymore.
answered Apr 20 at 16:32
LuckyPalLuckyPal
342110
342110
add a comment |
add a comment |
$begingroup$
Assume the significance level is $alpha$, which when talking about the null hypothesis, we are usually looking at 5% or 1% and so on.
In simple terms: p-value is the smallest $alpha$ at which we reject the null hypothesis.
So, when your p-value is 0.15, then we accept the null hypothesis when $alpha$ is 5% (or our confidence interval is 90%). But change that to only have a confidence interval of 60% and you reject your null hypothesis. Similarly, when your p-value=0.0038, it means you accept the null hypothesis under any value smaller than < 0.38%. That's why you compare p-value with $alpha$ and if p-value < $alpha$, you say that you cannot accept the null hypothesis.
New contributor
$endgroup$
add a comment |
$begingroup$
Assume the significance level is $alpha$, which when talking about the null hypothesis, we are usually looking at 5% or 1% and so on.
In simple terms: p-value is the smallest $alpha$ at which we reject the null hypothesis.
So, when your p-value is 0.15, then we accept the null hypothesis when $alpha$ is 5% (or our confidence interval is 90%). But change that to only have a confidence interval of 60% and you reject your null hypothesis. Similarly, when your p-value=0.0038, it means you accept the null hypothesis under any value smaller than < 0.38%. That's why you compare p-value with $alpha$ and if p-value < $alpha$, you say that you cannot accept the null hypothesis.
New contributor
$endgroup$
add a comment |
$begingroup$
Assume the significance level is $alpha$, which when talking about the null hypothesis, we are usually looking at 5% or 1% and so on.
In simple terms: p-value is the smallest $alpha$ at which we reject the null hypothesis.
So, when your p-value is 0.15, then we accept the null hypothesis when $alpha$ is 5% (or our confidence interval is 90%). But change that to only have a confidence interval of 60% and you reject your null hypothesis. Similarly, when your p-value=0.0038, it means you accept the null hypothesis under any value smaller than < 0.38%. That's why you compare p-value with $alpha$ and if p-value < $alpha$, you say that you cannot accept the null hypothesis.
New contributor
$endgroup$
Assume the significance level is $alpha$, which when talking about the null hypothesis, we are usually looking at 5% or 1% and so on.
In simple terms: p-value is the smallest $alpha$ at which we reject the null hypothesis.
So, when your p-value is 0.15, then we accept the null hypothesis when $alpha$ is 5% (or our confidence interval is 90%). But change that to only have a confidence interval of 60% and you reject your null hypothesis. Similarly, when your p-value=0.0038, it means you accept the null hypothesis under any value smaller than < 0.38%. That's why you compare p-value with $alpha$ and if p-value < $alpha$, you say that you cannot accept the null hypothesis.
New contributor
New contributor
answered Apr 20 at 21:54
EhsanKEhsanK
1163
1163
New contributor
New contributor
add a comment |
add a comment |
Rohit Saluja is a new contributor. Be nice, and check out our Code of Conduct.
Rohit Saluja is a new contributor. Be nice, and check out our Code of Conduct.
Rohit Saluja is a new contributor. Be nice, and check out our Code of Conduct.
Rohit Saluja is a new contributor. Be nice, and check out our Code of Conduct.
Thanks for contributing an answer to Cross Validated!
- Please be sure to answer the question. Provide details and share your research!
But avoid …
- Asking for help, clarification, or responding to other answers.
- Making statements based on opinion; back them up with references or personal experience.
Use MathJax to format equations. MathJax reference.
To learn more, see our tips on writing great answers.
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function ()
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstats.stackexchange.com%2fquestions%2f404119%2funderstanding-p-values-using-an-example%23new-answer', 'question_page');
);
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function ()
StackExchange.helpers.onClickDraftSave('#login-link');
);
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
$begingroup$
You say the population standard deviation $sigma$ is known. Can you provide the known value? The terminology "correction factor" is not familiar to me; can you give a formula for finding that? // The sample mean $bar X = 2.92$ is below the hypothetical population mean $mu_0 = 3.$ The issue is whether it is enough smaller to warrant rejecting the null hypothesis.
$endgroup$
– BruceET
Apr 20 at 13:20
$begingroup$
The population standard deviation is .18 and sample size is 36, hence the correction factor is 0.18/sqrt(36) equals 0.03
$endgroup$
– Rohit Saluja
Apr 20 at 13:24
2
$begingroup$
Thanks for the additional information. The usual terminology is to call $sigma/sqrtn$ the 'standard error'.
$endgroup$
– BruceET
Apr 20 at 13:27
$begingroup$
@BruceET - the issue is if it is enough smaller to reject the null hypotheses, however the probability of sample mean of being less than or equal to 2.92 is only 0.0038, so can't we say that the probability of sample mean of less than 3 is very less hence we support null hypotheses.
$endgroup$
– Rohit Saluja
Apr 20 at 13:39