{"id":79,"date":"2019-06-11T19:16:38","date_gmt":"2019-06-11T19:16:38","guid":{"rendered":"http:\/\/blogs.ubalt.edu\/jboettinger\/?p=79"},"modified":"2019-10-17T20:27:37","modified_gmt":"2019-10-17T20:27:37","slug":"statistics-introduction-to-the-t-statistic","status":"publish","type":"post","link":"https:\/\/blogs.ubalt.edu\/mathsupportcenter\/2019\/06\/11\/statistics-introduction-to-the-t-statistic\/","title":{"rendered":"Statistics: Introduction to the t-statistic"},"content":{"rendered":"<h1><b>Introduction to the t-statistic<\/b><\/h1>\n<h2><strong>Z-tests vs. t-tests<\/strong><\/h2>\n<p><strong>Z-tests<\/strong><span style=\"font-weight: 400\">\u00a0compare the means between a population and a sample and require information that is usually unavailable about populations, namely the variance\/standard deviation.\u00a0<\/span><span style=\"font-weight: 400\"><strong>Single sample t-tests<\/strong> compare the population mean to a sample mean, but only require one variance\/standard deviation, and that\u2019s from the sample. This is where\u00a0<strong>estimated standard error\u00a0<\/strong>comes in. It&#8217;s us<\/span><span style=\"font-weight: 400\">ed as an estimate of the real standard error, \u03c3<\/span><span style=\"font-weight: 400\">M<\/span><span style=\"font-weight: 400\">, when the value of \u03c3 is unknown. It is computed using the sample variance or sample standard deviation and provides an estimate of the standard distance between a sample mean, M, and the population mean, \u03bc, (or rather, the mean of sample means). It\u2019s an \u201cerror\u201d because it\u2019s the distance between what the sample mean is and what it would ideally be since we would rather have the population standard deviation. The formula for estimated standard error is\u00a0s\/\u221an.<\/span><\/p>\n<p>The formula for the t-test itself is:\u00a0<img loading=\"lazy\" decoding=\"async\" class=\"alignnone wp-image-100\" src=\"http:\/\/blogs.ubalt.edu\/jboettinger\/wp-content\/uploads\/sites\/1114\/2019\/06\/ttest.jpg\" alt=\"\" width=\"134\" height=\"93\" \/>\u00a0 with the bottom portion referring to the estimated standard error. You may see this written as sM instead.<!--more--><\/p>\n<h2>Degrees of Freedom<\/h2>\n<p><span style=\"font-weight: 400\"><strong>Degrees of freedom<\/strong> describe the number of scores in a sample that are independent and free to vary. Because the sample mean places a restriction on the value of one score in the sample, there are degrees of freedom for a sample with n scores. For a single sample t-test, the degrees of freedom are calculated using the following formula: df = n-1.<\/span><\/p>\n<h2>Shape of a t-distribution<\/h2>\n<p>One general rule of t-distribution is that it&#8217;s always slightly flatter than it&#8217;s corresponding normal distribution. This is because t-statistics are always working from a sample size, which is relatively small, rather than a population, which are generally large.<\/p>\n<p>There are some factors which also influence the shape of each individual t-distribution:<\/p>\n<ul>\n<li><span style=\"font-weight: 400\"><span style=\"text-decoration: underline\">The degrees of freedom<\/span>: The greater n is, and subsequently df, the more like a normal distribution the t-distribution will begin to look (this also follows the 30 rule)<\/span><\/li>\n<li><span style=\"font-weight: 400\"><span style=\"text-decoration: underline\">The sample variance<\/span>: The bottom half of the equation deals with the estimated standard error, which changes when the standard deviation changes. Because the ESE is dependent on the sample variance, each sample can create a different ESE<\/span><\/li>\n<\/ul>\n<h2>Hypothesis Testing with a Single Sample t-test<\/h2>\n<p><span style=\"font-weight: 400\">The written null and alternative hypotheses for a single sample t-test are as follows:<\/span><\/p>\n<p><span style=\"font-weight: 400\">H<\/span><span style=\"font-weight: 400\">0<\/span><i><span style=\"font-weight: 400\"> : \u03bc = population mean<\/span><\/i><\/p>\n<p><span style=\"font-weight: 400\">H<\/span><span style=\"font-weight: 400\">1<\/span><span style=\"font-weight: 400\"> : \u03bc \u2260 population mean<\/span><\/p>\n<p><span style=\"font-weight: 400\">The stops for a single sample t-test are as follows:\u00a0<\/span><\/p>\n<ol>\n<li><span style=\"font-weight: 400\">State the hypotheses and select an alpha level<\/span><\/li>\n<li><span style=\"font-weight: 400\">Locate the critical region<\/span><\/li>\n<li><span style=\"font-weight: 400\">Calculate the test statistic<\/span><\/li>\n<li><span style=\"font-weight: 400\">Make a decision regarding the null hypothesis<\/span><\/li>\n<\/ol>\n<p>The following are some assumptions one makes when doing a single sample t-test:<\/p>\n<ul>\n<li style=\"font-weight: 400\"><span style=\"font-weight: 400\">The values in the sample must consist of independent observations.<\/span><\/li>\n<li style=\"font-weight: 400\"><span style=\"font-weight: 400\">The population that is sampled must be normal.<\/span><\/li>\n<\/ul>\n<h2>Effect Size for Single Sample t-tests<\/h2>\n<p>Effect size for a single sample t-test is calculated using Cohen&#8217;s d. The formula for this is the mean difference over the standard deviation, or\u00a0<span style=\"font-weight: 400\">(M \u2013 \u03bc) \/ s.\u00a0<\/span><span style=\"font-weight: 400\">Effect size is important because it\u2019s a way of quantifying the difference between two groups, rather than just saying that there is a significant difference. Essentially, it\u2019s also important to know how much of a difference there is, not just the likelihood that the group differences you\u2019re seeing are a fluke because the null hypothesis is actually true. This can also be important for practical significance. If your treatment is statistically significant, but has a small effect size, is it worth using this treatment on clients?<\/span><\/p>\n<p><span style=\"font-weight: 400\">For Cohen\u2019s d, 0.2 would be considered a small effect size, 0.5 is medium, and 0.8 is large. Some people will mix words together like \u201csmall to medium effect size\u201d but some professors will want you to just pick a side.\u00a0<\/span><\/p>\n<h2>Confidence Intervals<\/h2>\n<p><span style=\"font-weight: 400\"><strong>Confidence intervals<\/strong> are a range of values which is likely to encompass the true value you\u2019re looking for. More specifically, it&#8217;s a range we create using a sample that we can say with X% confidence that the population mean falls within that range.\u00a0<\/span><span style=\"font-weight: 400\">Confidence intervals are constructed at a confidence level, such as 95%, selected by the user. It means that if the same population is sampled on numerous occasions and interval estimates are made on each occasion, the resulting intervals would bracket the true population parameter in approximately 95% of the cases.\u00a0<\/span><span style=\"font-weight: 400\">Confidence intervals and any kind of interval estimation are used in the same situations that you would use hypothesis testing. There is an estimation procedure for every kind of hypothesis test.<\/span><\/p>\n","protected":false},"excerpt":{"rendered":"<p>Introduction to the t-statistic Z-tests vs. t-tests Z-tests\u00a0compare the means between a population and a sample and require information that is usually unavailable about populations, namely the variance\/standard deviation.\u00a0Single sample t-tests compare the population mean to a sample mean, but only require one variance\/standard deviation, and that\u2019s from the sample. This is where\u00a0estimated standard error\u00a0comes [&hellip;]<\/p>\n","protected":false},"author":1347,"featured_media":0,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[151,153,148,152,158,155,156,150,159,160],"tags":[4,3,5,70,43,73,57,8,7,6,71,72],"_links":{"self":[{"href":"https:\/\/blogs.ubalt.edu\/mathsupportcenter\/wp-json\/wp\/v2\/posts\/79"}],"collection":[{"href":"https:\/\/blogs.ubalt.edu\/mathsupportcenter\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/blogs.ubalt.edu\/mathsupportcenter\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/blogs.ubalt.edu\/mathsupportcenter\/wp-json\/wp\/v2\/users\/1347"}],"replies":[{"embeddable":true,"href":"https:\/\/blogs.ubalt.edu\/mathsupportcenter\/wp-json\/wp\/v2\/comments?post=79"}],"version-history":[{"count":3,"href":"https:\/\/blogs.ubalt.edu\/mathsupportcenter\/wp-json\/wp\/v2\/posts\/79\/revisions"}],"predecessor-version":[{"id":120,"href":"https:\/\/blogs.ubalt.edu\/mathsupportcenter\/wp-json\/wp\/v2\/posts\/79\/revisions\/120"}],"wp:attachment":[{"href":"https:\/\/blogs.ubalt.edu\/mathsupportcenter\/wp-json\/wp\/v2\/media?parent=79"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/blogs.ubalt.edu\/mathsupportcenter\/wp-json\/wp\/v2\/categories?post=79"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/blogs.ubalt.edu\/mathsupportcenter\/wp-json\/wp\/v2\/tags?post=79"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}