{"id":73,"date":"2019-06-11T16:24:31","date_gmt":"2019-06-11T16:24:31","guid":{"rendered":"http:\/\/blogs.ubalt.edu\/jboettinger\/?p=73"},"modified":"2019-10-17T20:27:46","modified_gmt":"2019-10-17T20:27:46","slug":"statistics-introduction-to-hypothesis-testing","status":"publish","type":"post","link":"https:\/\/blogs.ubalt.edu\/mathsupportcenter\/2019\/06\/11\/statistics-introduction-to-hypothesis-testing\/","title":{"rendered":"Statistics: Introduction to Hypothesis Testing"},"content":{"rendered":"<h1>Hypothesis Testing Basics<\/h1>\n<h2><strong>What is Hypothesis Testing?<\/strong><\/h2>\n<p><span style=\"font-weight: 400\"><strong>Hypothesis testing<\/strong> is a big part of what we would actually consider testing for inferential statistics. It&#8217;s a procedure and set of rules that allow us to move from descriptive statistics to make inferences about a population based on sample data. It is a\u00a0<\/span><span style=\"font-weight: 400\">statistical method that uses sample data to evaluate a hypothesis about a population.<\/span><\/p>\n<p><span style=\"font-weight: 400\">This type of test is usually used within the context of research. If we expect to see a difference between a treated and untreated group (in some cases the untreated group is the parameters we know about the population), we expect there to be a difference in the means between the two groups, but that the standard deviation remains the same, as if each individual score has had a value added or subtracted from it.<\/span><!--more--><\/p>\n<h2><strong>Steps of Hypothesis Testing<\/strong><\/h2>\n<p>The following steps will be tailored to fit the first kind of hypothesis testing we will learn first: single-sample z-tests. There are many other kinds of tests, so keep this in mind.<\/p>\n<ul>\n<li><span style=\"font-weight: 400\">Step 1: State the Hypothesis<\/span>\n<ul>\n<li><b style=\"font-size: 1rem\">Null Hypothesis (H<\/b><b style=\"font-size: 1rem\">0):<\/b><span style=\"font-weight: 400\">\u00a0states that in the general population there is no change, no difference, or no relationship, or in the context of an experiment, it predicts that the independent variable has no effect on the dependent variable.<\/span><\/li>\n<li><b style=\"font-size: 1rem\">Alternative Hypothesis (H<\/b><b style=\"font-size: 1rem\">1):<\/b><span style=\"font-weight: 400\">\u00a0states that there is a change, a difference, or a relationship for the general population, or in the context of an experiment, it predicts that the independent variable has an effect on the dependent variable.<\/span><\/li>\n<\/ul>\n<\/li>\n<li>Step 2: Set the Criteria for a Decision\n<ul>\n<li><b style=\"font-size: 1rem\">Alpha Level<\/b><span style=\"font-weight: 400\">: Also known as <\/span><b style=\"font-size: 1rem\">Level of Significance<\/b><span style=\"font-weight: 400\">, is a probability value that is used to define the concept of \u201cvery unlikely\u201d in a hypothesis test. We chose an alpha level in order to separate the most unlikely sample means from the most likely sample means. Ex. \u03b1 = 0.05, that means that we\u2019re separating the most unlikely 5% from the most likely 95%.\u00a0<\/span>The largest permissible value of alpha is 0.05, although some researchers like to use more conservative alpha levels to reduce the risk that a false report is published. But you don\u2019t want the value to be too conservative because otherwise, you might run the risk of a Type II error, in which case the hypothesis test demands more evidence from the research results, in which case you might be throwing out evidence that a treatment will work.<\/li>\n<li><b style=\"font-size: 1rem\">Critical Region<\/b><span style=\"font-weight: 400\">: Composed of the extreme sample values that are very unlikely to be obtained if the null hypothesis is true. Determined by alpha level. If sample data fall in the critical region, the null hypothesis is rejected, because it\u2019s very unlikely they\u2019ve fallen there by chance.<\/span><\/li>\n<\/ul>\n<\/li>\n<li>Step 3: Collect Data and Compute Sample Statistics\n<ul>\n<li>After collecting the data, we find the sample mean. Now we can compare the sample mean with the null hypothesis by computing a z-score that describes where the sample mean is located relative to the hypothesized population mean. We use the z-score formula.<\/li>\n<li>\u00a0Step 4: Make a Decision\n<ul>\n<li>We decided previously what the two z-score boundaries are for a critical score. If the z-score we get after plugging the numbers in the aforementioned equation is outside of that critical region, we reject the null hypothesis. Otherwise, we would say that we failed to reject the null hypothesis.<\/li>\n<\/ul>\n<\/li>\n<\/ul>\n<\/li>\n<\/ul>\n<h2>Regions of the Distribution<\/h2>\n<p><span style=\"font-weight: 400\">Because we&#8217;re making judgments based on probability and proportion, our normal distributions and certain regions within them come into play.<\/span><\/p>\n<p>As mentioned before,<strong> Alpha<\/strong><b> Level<\/b>,\u00a0a<span style=\"font-weight: 400\">lso known as <\/span><b>Level of Significance<\/b><span style=\"font-weight: 400\">, is a probability value that is used to define the concept of \u201cvery unlikely\u201d in a hypothesis test. We chose an alpha level in order to separate the most unlikely sample means from the most likely sample means. Ex. \u03b1 = 0.05, that means that we\u2019re separating the most unlikely 5% from the most likely 95%<\/span><\/p>\n<p>The <strong>Critical<\/strong><b> Region\u00a0<\/b><span style=\"font-weight: 400\">is composed of the extreme sample values that are very unlikely to be obtained if the null hypothesis is true. Determined by alpha level. If sample data fall in the critical region, the null hypothesis is rejected, because it\u2019s very unlikely they\u2019ve fallen there by chance.<\/span><\/p>\n<p><span style=\"font-weight: 400\">These regions come into play when talking about different errors.<\/span><\/p>\n<p><span style=\"font-weight: 400\">A <\/span><b>Type I Error<\/b><span style=\"font-weight: 400\"> occurs when a researcher rejects a null hypothesis that is actually true; the researcher concludes that a treatment has an effect when it actually doesn\u2019t. This happens when a researcher unknowingly obtains an extreme, nonrepresentative sample. This goes back to alpha level: it\u2019s the probability that the test will lead to a Type I error if the null hypothesis is true.<\/span><\/p>\n<p><span style=\"font-weight: 400\">A <\/span><b>Type II Error<\/b><span style=\"font-weight: 400\"> occurs when a researcher fails to reject the null hypothesis that is really false; this means that the hypothesis test has failed to detect a real treatment effect. This happens when the sample mean is not in the critical region even though the treatment has had an effect on the sample. Usually, this means that the effect of the treatment was small, but it\u2019s still there. The probability of a Type II error is represented by beta (\u03b2)<\/span><\/p>\n<p>&nbsp;<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone wp-image-74\" src=\"http:\/\/blogs.ubalt.edu\/jboettinger\/wp-content\/uploads\/sites\/1114\/2019\/06\/type1type2-300x212.gif\" alt=\"\" width=\"524\" height=\"370\" \/><\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"alignnone wp-image-75\" src=\"http:\/\/blogs.ubalt.edu\/jboettinger\/wp-content\/uploads\/sites\/1114\/2019\/06\/error-and-power-3-728-300x225.jpg\" alt=\"\" width=\"460\" height=\"345\" srcset=\"https:\/\/blogs.ubalt.edu\/mathsupportcenter\/wp-content\/uploads\/sites\/1114\/2019\/06\/error-and-power-3-728-300x225.jpg 300w, https:\/\/blogs.ubalt.edu\/mathsupportcenter\/wp-content\/uploads\/sites\/1114\/2019\/06\/error-and-power-3-728-624x468.jpg 624w, https:\/\/blogs.ubalt.edu\/mathsupportcenter\/wp-content\/uploads\/sites\/1114\/2019\/06\/error-and-power-3-728.jpg 728w\" sizes=\"(max-width: 460px) 100vw, 460px\" \/><\/p>\n<p><em>Source:\u00a0https:\/\/www.slideshare.net\/shoffma5\/error-and-power<\/em><\/p>\n<p>A result is said to be <strong>significant<\/strong> or <strong>statistically significant<\/strong> if it is very unlikely to occur when the null hypothesis is true. That is, the result is sufficient to reject the null hypothesis. For instance, two means can be significantly different from one another.<\/p>\n<h2>Factors that Influence and Assumptions of Hypothesis Testing<\/h2>\n<p>Assumptions of Hypothesis Testing:<\/p>\n<ul>\n<li><strong>Random sampling: <\/strong>it is assumed that the participants used in the study were selected randomly so that we can confidently generalize our findings from the sample to the population.<\/li>\n<li><strong style=\"font-size: 1rem\">Independent observation: <\/strong>two observations are independent if there is no consistent, predictable relationship between the first observation and the second.<\/li>\n<li>The value of \u03c3 is unchanged by the treatment; if the population standard deviation is unknown, we assume that the standard deviation for the unknown population (after treatment) is the same as it was for the population before treatment. There are ways of checking to see if this is true in SPSS or Excel.<\/li>\n<li><strong>Normal sampling distribution<\/strong>: in order to use the unit normal table to identify the critical region, we need the distribution of sample means to be normal (which means we need the population to be distributed normally and\/or each sample size needs to be 30 or greater based on what we know about the central limit theorem).<\/li>\n<\/ul>\n<p>Factors that influence hypothesis testing:<\/p>\n<ul>\n<li>The variability of the scores, which is measured by either the standard deviation or the variance. The variability influences the size of the standard error in the denominator of the z-score.<\/li>\n<li>The number of scores in the sample. This value also influences the size of the standard error in the denominator.<\/li>\n<\/ul>\n<p><strong>Test statistic: <\/strong>indicates that the sample data are converted into a single, specific statistic that is used to test the hypothesis (in this case, the z-score statistic).<\/p>\n<h2>Directional Hypotheses and Tailed Tests<\/h2>\n<p>In a <strong>directional hypothesis test<\/strong>, also known as a one-tailed test, the statistical hypotheses specify with an increase or decrease in the population mean. That is, they make a statement about the direction of the effect.<\/p>\n<p>The Hypotheses for a Directional Test:<\/p>\n<ul>\n<li>H0: The test scores are not increased\/decreased (the treatment doesn\u2019t work)<\/li>\n<li>H<span style=\"font-size: 1rem\">1<\/span><span style=\"font-size: 1rem\">: The test scores are increased\/decreased (the treatment works as predicted)<\/span><\/li>\n<\/ul>\n<p>Because we\u2019re only worried about scores that are either greater or less than the scores predicted by the null hypothesis, we only worry about what\u2019s going on in one tail meaning that the critical region only exists within one tail.\u00a0This means that all of the alpha is contained in one tail rather than split up into both (so the whole 5% is located in the tail we care about, rather than 2.5% in each tail).\u00a0So before, we cared about what\u2019s going on at the 0.025 mark of the unit normal table to look at both tails, but now we care about 0.05 because we\u2019re only looking at one tail.<\/p>\n<p>A one-tailed test allows you to reject the null hypothesis when the difference between the sample and the population is relatively small, as long as that difference is in the direction that you predicted. A two-tailed test, on the other hand, requires a relatively large difference independent of direction. In practice, researchers hypothesize using a one-tailed method but base their findings off of whether the results fall into the critical region of a two-tailed method. For the purposes of this class, make sure to calculate your results using the test that is specified in the problem.<\/p>\n<h2><strong>Effect Size<\/strong><\/h2>\n<p>A measure of <strong>effect size<\/strong> is intended to provide a measurement of the absolute magnitude of a treatment effect, independent of the size of the sample(s) being used. Usually done with Cohen\u2019s d.\u00a0If you imagine the two distributions, they\u2019re layered over one another. The more they overlap, the smaller the effect size (the means of the two distributions are close). The more they are spread apart, the greater the effect size (the means of the two distributions are farther apart).<\/p>\n<h2>Statistical Power<\/h2>\n<p>The <strong>power<\/strong> of a statistical test is the probability that the test will correctly reject a false null hypothesis. It&#8217;s usually what we&#8217;re hoping to get when we run an experiment. It&#8217;s displayed in the table posted above.\u00a0Power and effect size are connected. So, we know that the greater the distance between the means, the greater the effect size. If the two distributions overlapped very little, there would be a greater chance of selecting a sample that leads to rejecting the null hypothesis.<\/p>\n","protected":false},"excerpt":{"rendered":"<p>Hypothesis Testing Basics What is Hypothesis Testing? Hypothesis testing is a big part of what we would actually consider testing for inferential statistics. It&#8217;s a procedure and set of rules that allow us to move from descriptive statistics to make inferences about a population based on sample data. It is a\u00a0statistical method that uses sample [&hellip;]<\/p>\n","protected":false},"author":1347,"featured_media":0,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[151,153,148,152,158,155,156,150,159,160],"tags":[4,3,5,60,59,62,68,57,61,8,58,69,7,6,65,66,67,63,64],"_links":{"self":[{"href":"https:\/\/blogs.ubalt.edu\/mathsupportcenter\/wp-json\/wp\/v2\/posts\/73"}],"collection":[{"href":"https:\/\/blogs.ubalt.edu\/mathsupportcenter\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/blogs.ubalt.edu\/mathsupportcenter\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/blogs.ubalt.edu\/mathsupportcenter\/wp-json\/wp\/v2\/users\/1347"}],"replies":[{"embeddable":true,"href":"https:\/\/blogs.ubalt.edu\/mathsupportcenter\/wp-json\/wp\/v2\/comments?post=73"}],"version-history":[{"count":3,"href":"https:\/\/blogs.ubalt.edu\/mathsupportcenter\/wp-json\/wp\/v2\/posts\/73\/revisions"}],"predecessor-version":[{"id":119,"href":"https:\/\/blogs.ubalt.edu\/mathsupportcenter\/wp-json\/wp\/v2\/posts\/73\/revisions\/119"}],"wp:attachment":[{"href":"https:\/\/blogs.ubalt.edu\/mathsupportcenter\/wp-json\/wp\/v2\/media?parent=73"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/blogs.ubalt.edu\/mathsupportcenter\/wp-json\/wp\/v2\/categories?post=73"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/blogs.ubalt.edu\/mathsupportcenter\/wp-json\/wp\/v2\/tags?post=73"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}