<img src="https://d5nxst8fruw4z.cloudfront.net/atrk.gif?account=iA1Pi1a8Dy00ym" style="display:none" height="1" width="1" alt="" />

# 7.6: Student’s t-Distribution

Difficulty Level: At Grade Created by: CK-12

## Learning Objectives

• Use Student’s \begin{align*}t\end{align*}-distribution to estimate population mean interval for smaller samples.
• Understand how the shape of Student’s \begin{align*}t\end{align*}-distribution corresponds to the sample size (which corresponds to a measure called the “degrees of freedom.”)

## Introduction

In a previous lesson you learned about the Central Limit Theorem. One of the attributes of this theorem was that the sampling distribution of sample mean will follow a normal distribution as long as the sample size is large. As the value of \begin{align*}n\end{align*} increases, the sampling distribution is more and more likely to follow a normal distribution. You’ve also learned that when the standard deviation of a population is known, a \begin{align*}z\end{align*}-score can be calculated and used with the normal distribution to evaluate probabilities with the sample mean. In real-life situations, the standard deviation of the entire population \begin{align*}(\sigma)\end{align*}, is rarely known. Also the sample size is not always large enough to emulate a normal distribution. In fact there are often times when the sample sizes are quite small. What do you do when either one or both of these events occur?

## t-Statistic

People often make decisions from data by comparing the results from a sample to some hypothesized or predetermined parameter. These decisions are referred to as tests of significance or hypothesis tests since they are used to determine whether the predetermined parameter is acceptable or should be rejected. We know that if we flip a fair coin, the probability of getting heads is \begin{align*}0.5\end{align*}. In other words, heads and tails are equally likely. Therefore, when a coin is spun, it should land heads \begin{align*}50\%\end{align*} of the time. Let’s say that a coin of questionable fairness was spun \begin{align*}40\;\mathrm{times}\end{align*} it landed heads \begin{align*}12 \;\mathrm{times}\end{align*}. For these spins the sample proportion of heads is \begin{align*}\hat {p} = \frac{12} {40} = 0.3\end{align*}. If technology is used to determine a \begin{align*}95\%\end{align*} confidence interval to support the standard that heads should land \begin{align*}50\%\end{align*} of the time, the reasonably likely sample proportions are in the interval \begin{align*}0.34505\end{align*} to \begin{align*}0.65495\end{align*}. The class with \begin{align*}\hat {p} = 0.3\end{align*}, is not captured within this confidence interval. Therefore, the fairness of this coin should be questioned; or, in other words, value of \begin{align*}0.5\end{align*} as a plausible value for the proportion of times this particular coin lands heads when it is spun should be rejected. This data has provided evidence against the standard.

The object is to test the significance of the difference between the sample and the parameter. If the difference is small (as defined by some predetermined amount), then the parameter is acceptable. The statement that the proposed parameter is true is called the null hypothesis. If the difference is large and can’t reasonably be attributed to chance, then the parameter can be rejected.

When the sample size is large, reliable estimates of the mean and variance of the population from which the sample was drawn can be made. Up to this point, we have used the \begin{align*}z\end{align*}-score to determine the number of standard deviations a given value lays above or below the mean.

\begin{align*}z = \frac{\bar {x} - \mu_0} {\sigma / \sqrt{n}}\end{align*}

where \begin{align*}\bar {x}\end{align*} is the sample mean, \begin{align*}\mu_0\end{align*} is the hypothesized mean stated in the null hypothesis \begin{align*}H_0: \mu= \mu_0 \ \sigma\end{align*}, is the population standard deviation and \begin{align*}n\end{align*} is the sample size.

However the above formula cannot be used to determine how far a sample mean is from the hypothesized mean because the standard deviation of the population is not known. If the value of \begin{align*}\sigma\end{align*} is unknown, \begin{align*}s\end{align*} is substituted for \begin{align*}\sigma\end{align*} and \begin{align*}t\end{align*} for \begin{align*}z\end{align*}. The \begin{align*}t\end{align*} stands for the “test statistic,” and it is given by the formula:

\begin{align*}t = \frac{\bar {x} - \mu_0} {s / \sqrt{n}}\end{align*}

where \begin{align*}\bar {x}\end{align*} is the sample mean \begin{align*}\mu_0\end{align*} is the population mean, \begin{align*}s\end{align*} is the standard deviation of the sample and \begin{align*}n\end{align*} is the sample size. The population mean \begin{align*}\mu\end{align*} is unknown but an estimate for this value is used. The \begin{align*}t\end{align*}-test will be used to determine the difference between the sample mean and the hypothesized mean. The null hypothesis that is being tested is \begin{align*}H_0:\mu = \mu_0\end{align*}

So, suppose you want to see if a hypothesized mean passes a \begin{align*}95\%\end{align*} level of confidence. The corresponding confidence interval can be determined by using the graphing calculator:

Press ENTER

\begin{align*}x =\end{align*} the number of successes in the sample and

\begin{align*}n =\end{align*} the sample size

Press ENTER again. The confidence level will appear on the next screen. The value for \begin{align*}t\end{align*} can now be compared with this interval to tell us whether or not the hypothesized mean can be accepted or rejected for this level of confidence.

Example:

The masses of newly produced bus tokens are estimated to have a mean of \begin{align*}3.16 \;\mathrm{grams}\end{align*}. A random sample of \begin{align*}11\end{align*} tokens was removed from the production line and the mean weight of the tokens was calculated as \begin{align*}3.21\;\mathrm{grams}\end{align*} with a standard deviation of \begin{align*}0.067\end{align*}. What is the value of the test statistic for a test to determine how the mean differs from the estimated mean?

Solution:

\begin{align*}t & = \frac{\bar {x} - \mu} {s/ \sqrt{n}} \\ t & = \frac{3.21 - 3.16} {0.067/ \sqrt{11}} \\ t & \approx 2.48\end{align*}

If the value of \begin{align*}t\end{align*} from the sample fits right into the middle of the distribution of \begin{align*}t\end{align*} constructed by assuming the null hypothesis is true, the null hypothesis is true. On the other hand, if the value of \begin{align*}t\end{align*} from the sample is way out in the tail of the \begin{align*}t\end{align*} -distribution, then there is evidence to reject the null hypothesis. Now that the distribution of \begin{align*}t\end{align*} is known when the null hypothesis is true, the location of this value on the distribution. The most common method used to determine this is to find a \begin{align*}P\end{align*}-value (observed significance level). The \begin{align*}P\end{align*}-value is a probability that is computed with the assumption that the null hypothesis is true.

The \begin{align*}P\end{align*}-value for a two-sided test is the area under the \begin{align*}t\end{align*}-distribution with \begin{align*}df = 11 -1\end{align*}, or \begin{align*}10\end{align*}, that lies above \begin{align*}t = 2.48\end{align*} and below \begin{align*}t = -2.48\end{align*}. This \begin{align*}P\end{align*}-value can be calculated by using technology.

Press 2ND [DIST] Use \begin{align*}\downarrow\end{align*} to select 5.tcdf (lower bound, upper bound, degrees of freedom)

This will be the total area under both tails. To calculate the area under one tail divide by \begin{align*}2\end{align*}.

There is only a \begin{align*}0.016\end{align*} chance of getting an absolute value of \begin{align*}t\end{align*} as large as or even larger than the one from this sample \begin{align*}(2.48 \le t \le -2.48)\end{align*}. The small \begin{align*}P\end{align*}-value tells us that the sample is inconsistent with the null hypothesis. The population mean differs from the estimated mean of \begin{align*}3.16\end{align*}.

When the \begin{align*}P\end{align*}-value is close to zero, there is strong evidence against the null hypothesis. When the \begin{align*}P\end{align*}-value is large, the result form the sample is consistent with the estimated or hypothesized mean and there is no evidence against the null hypothesis.

A visual picture of the \begin{align*}P\end{align*}-value can be obtained by using the graphing calculator.

This formula \begin{align*}t = \frac{\bar {x} - \mu} {s/ \sqrt{n}}\end{align*} is similar to that used in computing the \begin{align*}z\end{align*} statistic with the unknown population standard deviation \begin{align*}(\sigma)\end{align*} being substituted with the sample standard deviation.

There are numerous \begin{align*}t\end{align*}-distributions and all are determined by a property of a set of data called the number of degrees of freedom. The degrees of freedom refer to the number of independent observations in a set of data. When estimating a mean score from a single sample, the number of independent observations is equal to the sample size minus one. In a single sample, there are \begin{align*}n\end{align*} observations but only one parameter that needs to be estimated (the mean). This means that there are \begin{align*}n - 1\;\mathrm{degrees}\end{align*} of freedom for estimating variability. In other words \begin{align*}df = n - 1\end{align*}, where \begin{align*}n\end{align*} is the sample size. The distribution of the \begin{align*}t\end{align*}-statistic from samples of size \begin{align*}7\end{align*} would be described by a \begin{align*}t\end{align*}-distribution having \begin{align*}7 - 1\end{align*} or \begin{align*}6 \;\mathrm{degrees}\end{align*} of freedom. Likewise, a \begin{align*}t\end{align*}-distribution with \begin{align*}12\;\mathrm{degrees}\end{align*} of freedom would be used with a sample size of \begin{align*}13\end{align*}.

The \begin{align*}t\end{align*}-score produced by this formula is associated with a unique cumulative probability which represents the chance of finding a sample mean less than or equal to \begin{align*}\bar {x}\end{align*}, using a random sample of size \begin{align*}n\end{align*}. The symbol \begin{align*}t_\alpha\end{align*} is used to represent the \begin{align*}t\end{align*}-score that has a cumulative probability of \begin{align*}(1 - \alpha)\end{align*}. If you needed the \begin{align*}t\end{align*}-score to have a cumulative probability of \begin{align*}0.95\end{align*}, then \begin{align*}\alpha\end{align*} would be equal to \begin{align*}(1-0.95)\end{align*} or simply \begin{align*}0.05\end{align*}. This means that the \begin{align*}t\end{align*}-score would be written as \begin{align*}t_{0.05}\end{align*}. This value depends on the number of degrees of freedom and this value can be determined by using the table of values:

df\p \begin{align*}0.40\end{align*} \begin{align*}0.25\end{align*} \begin{align*}0.10\end{align*} \begin{align*}0.05\end{align*} \begin{align*}0.025\end{align*} \begin{align*}0.01\end{align*} \begin{align*}0.005\end{align*} \begin{align*}0.0005\end{align*}
1 \begin{align*}0.324920\end{align*} \begin{align*}1.000000\end{align*} \begin{align*}3.077684\end{align*} \begin{align*}6.313752\end{align*} \begin{align*}12.70620\end{align*} \begin{align*}31.82052\end{align*} \begin{align*}63.65674\end{align*} \begin{align*}636.6192\end{align*}
2 \begin{align*}0.288675\end{align*} \begin{align*}0.816497\end{align*} \begin{align*}1.885618\end{align*} \begin{align*}2.919986\end{align*} \begin{align*}4.30265\end{align*} \begin{align*}6.96456\end{align*} \begin{align*}9.92484\end{align*} \begin{align*}31.5991\end{align*}
3 \begin{align*}0.276671\end{align*} \begin{align*}0.764892\end{align*} \begin{align*}1.637744\end{align*} \begin{align*}2.353363\end{align*} \begin{align*}3.18245\end{align*} \begin{align*}4.54070\end{align*} \begin{align*}5.84091\end{align*} \begin{align*}12.9240\end{align*}
4 \begin{align*}0.270722\end{align*} \begin{align*}0.740697\end{align*} \begin{align*}1.533206\end{align*} \begin{align*}2.131847\end{align*} \begin{align*}2.77645\end{align*} \begin{align*}3.74695\end{align*} \begin{align*}4.60409\end{align*} \begin{align*}8.6103\end{align*}
5 \begin{align*}0.267181\end{align*} \begin{align*}0.726687\end{align*} \begin{align*}1.475884\end{align*} \begin{align*}2.015048\end{align*} \begin{align*}2.57058\end{align*} \begin{align*}3.36493\end{align*} \begin{align*}4.03214\end{align*} \begin{align*}6.8688\end{align*}
6 \begin{align*}0.264835\end{align*} \begin{align*}0.717558\end{align*} \begin{align*}1.439756\end{align*} \begin{align*}1.943180\end{align*} \begin{align*}2.44691\end{align*} \begin{align*}3.14267\end{align*} \begin{align*}3.70743\end{align*} \begin{align*}5.9588\end{align*}
7 \begin{align*}0.263167\end{align*} \begin{align*}0.711142\end{align*} \begin{align*}1.414924\end{align*} \begin{align*}1.894579\end{align*} \begin{align*}2.36462\end{align*} \begin{align*}2.99795\end{align*} \begin{align*}3.49948\end{align*} \begin{align*}5.4079\end{align*}
8 \begin{align*}0.261921\end{align*} \begin{align*}0.706387\end{align*} \begin{align*}1.396815\end{align*} \begin{align*}1.859548\end{align*} \begin{align*}2.30600\end{align*} \begin{align*}2.89646\end{align*} \begin{align*}3.35539\end{align*} \begin{align*}5.0413\end{align*}
9 \begin{align*}0.260955\end{align*} \begin{align*}0.702722\end{align*} \begin{align*}1.383029\end{align*} \begin{align*}1.833113\end{align*} \begin{align*}2.26216\end{align*} \begin{align*}2.82144\end{align*} \begin{align*}3.24984\end{align*} \begin{align*}4.7809\end{align*}
10 \begin{align*}0.260185\end{align*} \begin{align*}0.699812\end{align*} \begin{align*}1.372184\end{align*} \begin{align*}1.812461\end{align*} \begin{align*}2.22814\end{align*} \begin{align*}2.76377\end{align*} \begin{align*}3.16927\end{align*} \begin{align*}4.5869\end{align*}
11 \begin{align*}0.259556\end{align*} \begin{align*}0.697445\end{align*} \begin{align*}1.363430\end{align*} \begin{align*}1.795885\end{align*} \begin{align*}2.20099\end{align*} \begin{align*}2.71808\end{align*} \begin{align*}3.10581\end{align*} \begin{align*}4.4370\end{align*}
12 \begin{align*}0.259033\end{align*} \begin{align*}0.695483\end{align*} \begin{align*}1.356217\end{align*} \begin{align*}1.782288\end{align*} \begin{align*}2.17881\end{align*} \begin{align*}2.68100\end{align*} \begin{align*}3.05454\end{align*} \begin{align*}4.3178\end{align*}
13 \begin{align*}0.258591\end{align*} \begin{align*}0.693829\end{align*} \begin{align*}1.350171\end{align*} \begin{align*}1.770933\end{align*} \begin{align*}2.16037\end{align*} \begin{align*}2.65031\end{align*} \begin{align*}3.01228\end{align*} \begin{align*}4.2208\end{align*}
14 \begin{align*}0.258213\end{align*} \begin{align*}0.692417\end{align*} \begin{align*}1.345030\end{align*} \begin{align*}1.761310\end{align*} \begin{align*}2.14479\end{align*} \begin{align*}2.62449\end{align*} \begin{align*}2.97684\end{align*} \begin{align*}4.1405\end{align*}
15 \begin{align*}0.257885\end{align*} \begin{align*}0.691197\end{align*} \begin{align*}1.340606\end{align*} \begin{align*}1.753050\end{align*} \begin{align*}2.13145\end{align*} \begin{align*}2.60248\end{align*} \begin{align*}2.94671\end{align*} \begin{align*}4.0728\end{align*}
16 \begin{align*}0.257599\end{align*} \begin{align*}0.690132\end{align*} \begin{align*}1.336757\end{align*} \begin{align*}1.745884\end{align*} \begin{align*}2.11991\end{align*} \begin{align*}2.58349\end{align*} \begin{align*}2.92078\end{align*} \begin{align*}4.0150\end{align*}
17 \begin{align*}0.257347\end{align*} \begin{align*}0.689195\end{align*} \begin{align*}1.333379\end{align*} \begin{align*}1.739607\end{align*} \begin{align*}2.10982\end{align*} \begin{align*}2.56693\end{align*} \begin{align*}2.89823\end{align*} \begin{align*}3.9651\end{align*}
18 \begin{align*}0.257123\end{align*} \begin{align*}0.688364\end{align*} \begin{align*}1.330391\end{align*} \begin{align*}1.734064\end{align*} \begin{align*}2.10092\end{align*} \begin{align*}2.55238\end{align*} \begin{align*}2.87844\end{align*} \begin{align*}3.9216\end{align*}
19 \begin{align*}0.256923\end{align*} \begin{align*}0.687621\end{align*} \begin{align*}1.327728\end{align*} \begin{align*}1.729133\end{align*} \begin{align*}2.09302\end{align*} \begin{align*}2.53948\end{align*} \begin{align*}2.86093\end{align*} \begin{align*}3.8834\end{align*}
20 \begin{align*}0.256743\end{align*} \begin{align*}0.686954\end{align*} \begin{align*}1.325341\end{align*} \begin{align*}1.724718\end{align*} \begin{align*}2.08596\end{align*} \begin{align*}2.52798\end{align*} \begin{align*}2.84534\end{align*} \begin{align*}3.8495\end{align*}
21 \begin{align*}0.256580\end{align*} \begin{align*}0.686352\end{align*} \begin{align*}1.323188\end{align*} \begin{align*}1.720743\end{align*} \begin{align*}2.07961\end{align*} \begin{align*}2.51765\end{align*} \begin{align*}2.83136\end{align*} \begin{align*}3.8193\end{align*}
22 \begin{align*}0.256432\end{align*} \begin{align*}0.685805\end{align*} \begin{align*}1.321237\end{align*} \begin{align*}1.717144\end{align*} \begin{align*}2.07387\end{align*} \begin{align*}2.50832\end{align*} \begin{align*}2.81876\end{align*} \begin{align*}3.7921\end{align*}
23 \begin{align*}0.256297\end{align*} \begin{align*}0.685306\end{align*} \begin{align*}1.319460\end{align*} \begin{align*}1.713872\end{align*} \begin{align*}2.06866\end{align*} \begin{align*}2.49987\end{align*} \begin{align*}2.80734\end{align*} \begin{align*}3.7676\end{align*}
24 \begin{align*}0.256173\end{align*} \begin{align*}0.684850\end{align*} \begin{align*}1.317836\end{align*} \begin{align*}1.710882\end{align*} \begin{align*}2.06390\end{align*} \begin{align*}2.49216\end{align*} \begin{align*}2.79694\end{align*} \begin{align*}3.7454\end{align*}
25 \begin{align*}0.256060\end{align*} \begin{align*}0.684430\end{align*} \begin{align*}1.316345\end{align*} \begin{align*}1.708141\end{align*} \begin{align*}2.05954\end{align*} \begin{align*}2.48511\end{align*} \begin{align*}2.78744\end{align*} \begin{align*}3.7251\end{align*}
26 \begin{align*}0.255955\end{align*} \begin{align*}0.684043\end{align*} \begin{align*}1.314972\end{align*} \begin{align*}1.705618\end{align*} \begin{align*}2.05553\end{align*} \begin{align*}2.47863\end{align*} \begin{align*}2.77871\end{align*} \begin{align*}3.7066\end{align*}
27 \begin{align*}0.255858\end{align*} \begin{align*}0.683685\end{align*} \begin{align*}1.313703\end{align*} \begin{align*}1.703288\end{align*} \begin{align*}2.05183\end{align*} \begin{align*}2.47266\end{align*} \begin{align*}2.77068\end{align*} \begin{align*}3.6896\end{align*}
28 \begin{align*}0.255768\end{align*} \begin{align*}0.683353\end{align*} \begin{align*}1.312527\end{align*} \begin{align*}1.701131\end{align*} \begin{align*}2.04841\end{align*} \begin{align*}2.46714\end{align*} \begin{align*}2.76326\end{align*} \begin{align*}3.6739\end{align*}
29 \begin{align*}0.255684\end{align*} \begin{align*}0.683044\end{align*} \begin{align*}1.311434\end{align*} \begin{align*}1.699127\end{align*} \begin{align*}2.04523\end{align*} \begin{align*}2.46202\end{align*} \begin{align*}2.75639\end{align*} \begin{align*}3.6594\end{align*}
30 \begin{align*}0.255605\end{align*} \begin{align*}0.682756\end{align*} \begin{align*}1.310415\end{align*} \begin{align*}1.697261\end{align*} \begin{align*}2.04227\end{align*} \begin{align*}2.45726\end{align*} \begin{align*}2.75000\end{align*} \begin{align*}3.6460\end{align*}
inf \begin{align*}0.253347\end{align*} \begin{align*}0.674490\end{align*} \begin{align*}1.281552\end{align*} \begin{align*}1.644854\end{align*} \begin{align*}1.95996\end{align*} \begin{align*}2.32635\end{align*} \begin{align*}2.57583\end{align*} \begin{align*}3.2905\end{align*}

From the table it can be determined that \begin{align*}t_{0.05}\end{align*} for \begin{align*}2\;\mathrm{degrees}\end{align*} of freedom is \begin{align*}2.92\end{align*} while for \begin{align*}20 \;\mathrm{degrees}\end{align*} of freedom the value is \begin{align*}1.72\end{align*}.

Since the \begin{align*}t\end{align*}-distribution is symmetric about a mean of zero, the following statement is true.

\begin{align*}t_\alpha = -t_{1 - \alpha} && \text{and} && t_{1 - \alpha} = -t_\alpha\end{align*}

Therefore, if \begin{align*}t_{0.05} = 2.92\end{align*} then by applying the above statement \begin{align*}t_{0.95} = -2.92\end{align*}

A \begin{align*}t\end{align*}-distribution is mound shaped, with mean \begin{align*}0\end{align*} and a spread that depends on the degrees of freedom. The greater the degrees of freedom, the smaller the spread. As the number of degrees of freedom increases, the \begin{align*}t\end{align*}-distribution approaches a normal distribution. The spread of any \begin{align*}t\end{align*}-distribution is greater than that of a standard normal distribution. This is due to the fact that that in the denominator of the formula \begin{align*}\sigma\end{align*} has been replaced with \begin{align*}s\end{align*}. Since \begin{align*}s\end{align*} is a random quantity changing with various samples, the variability in \begin{align*}t\end{align*} is greater, resulting in a larger spread.

Notice in the first distribution graph the spread of the first (inner curve) is small but in the second one the both distributions are basically overlapping, so are roughly normal. This is due to the increase in the degrees of freedom.

Here are the \begin{align*}t\end{align*}-distributions for \begin{align*}df = 1\end{align*} and for \begin{align*}df = 12\end{align*} as graphed on the graphing calculator

You are now on the \begin{align*}Y =\end{align*} screen.

\begin{align*}Y = \mathrm{tpdf}( X, 1)\end{align*} [Graph]

Repeat the steps to plot more than one \begin{align*}t\end{align*}-distribution on the same screen.

Notice the difference in the two distributions.

The one with \begin{align*}12 = df\end{align*} approximates a normal curve.

The \begin{align*}t\end{align*}-distribution can be used with any statistic having a bell-shaped distribution. The Central Limit Theorem states the sampling distribution of a statistic will be close to normal with a large enough sample size. As a rough estimate, the Central Limit Theorem predicts a roughly normal distribution under the following conditions:

1. The population distribution is normal.
2. The sampling distribution is symmetric and the sample size is \begin{align*}\le 15\end{align*}.
3. The sampling distribution is moderately skewed and the sample size is \begin{align*}16 \le n \le 30\end{align*}.
4. The sample size is greater than \begin{align*}30\end{align*}, without outliers.

The \begin{align*}t\end{align*}-distribution also has some unique properties. These properties are:

1. The mean of the distribution equals zero.

2. The population standard deviation is unknown.

3. The variance is equal to the degrees of freedom divided by the degrees of freedom minus \begin{align*}2\end{align*}. This means that the degrees of freedom must be greater than two to avoid the expression being undefined.

\begin{align*}\text{Variance} = \frac{\text{df}} {\text{df} - 2}\ \text{and}\ \text{df} > 2\end{align*}

4. The variance is always greater than one, although it approaches \begin{align*}1\end{align*} as the degrees of freedom increase. This is due to the fact that as the degrees of freedom increase, the distribution is becoming more of a normal distribution.

5. Although the Student \begin{align*}t\end{align*}-distribution is bell-shaped, the smaller sample sizes produce a flatter curve. The distribution is not as mounded as a normal distribution and the tails are thicker. As the sample size increases and approaches \begin{align*}30\end{align*}, the distribution approaches a normal distribution.

6. The population is unimodal and symmetric.

Example:

Duracell manufactures batteries that the CEO claims will last \begin{align*}300\;\mathrm{hours}\end{align*} under normal use. A researcher randomly selected \begin{align*}15\end{align*} batteries from the production line and tested these batteries. The tested batteries had a mean life span of \begin{align*}290 \;\mathrm{hours}\end{align*} with a standard deviation of \begin{align*}50 \;\mathrm{hours}\end{align*}. If the CEO’s claim were true, what is the probability that \begin{align*}15\end{align*} randomly selected batteries would have a life span of no more than \begin{align*}290\;\mathrm{hours}\end{align*}?

Solution:

\begin{align*}t & = \frac{\bar {x} - \mu} {s/ \sqrt{n}} && \text{The degrees of freedom are } (n-1) = 15-1. \ \ \text{This means } 14\ \text{degrees of freedom}.\\ t & = \frac{290 - 300} {50 / \sqrt{15}}\\ t & = \frac{-10} {12.9099}\\ t & = -.7745993\end{align*}

Using the graphing calculator or a table of values, the cumulative probability is \begin{align*}0.286\end{align*}, which means that if the true life span of a battery were \begin{align*}300 \;\mathrm{hours}\end{align*}, there is a \begin{align*}28.6\%\end{align*} chance that the life span of the \begin{align*}15\end{align*} tested batteries would be less than or equal to \begin{align*}290\end{align*} days. This is not a high enough level of confidence to reject the null hypothesis and count the discrepancy as significant.

You are now on the \begin{align*}Y =\end{align*} screen.

\begin{align*}Y = \text{tpdf}( -.7745993, 14) = [0.286]\end{align*}

Example:

You have just taken ownership of a pizza shop. The previous owner told you that you would save money if you bought the mozzarella cheese in a \begin{align*}4.5\;\mathrm{pound}\end{align*} slab. Each time you purchase a slab of cheese, you weigh it to ensure that you are receiving \begin{align*}72 \;\mathrm{ounces}\end{align*} of cheese. The results of \begin{align*}7\end{align*} random measurements are \begin{align*}70, 69, 73, 68, 71, 69\end{align*} and \begin{align*}71\;\mathrm{ounces}\end{align*}. Are these differences due to chance or is the distributor giving you less cheese than you deserve?

Solution:

Begin the problem by determining the mean of the sample and the sample standard deviation.

This can be done using the graphing calculator. \begin{align*}\bar {x} \approx 70.143\end{align*} and \begin{align*}s \approx 1.676\end{align*}.

\begin{align*}t & = \frac{\bar {x} - \mu} {s/ \sqrt{n}}\\ t & = \frac{70.143 - 72} {1.676 / \sqrt{7}}\\ t & \approx -2.9315\end{align*}

Example:

In the example before last the test statistic for testing that the mean weight of the cheese wasn’t \begin{align*}72\end{align*} was computed. Find and interpret the \begin{align*}P\end{align*}-value.

Solution:

The test statistic computed in the example before last was \begin{align*}-2.9315\end{align*}. Using technology, the \begin{align*}P –\end{align*} value is \begin{align*}0.0262\end{align*}. If the mean weight of cheese is \begin{align*}72\;\mathrm{ounces}\end{align*}, the probability that the volume of \begin{align*}7\end{align*} random measurements would give a value of \begin{align*}t\end{align*} greater than \begin{align*}2.9315\end{align*} or less than \begin{align*}-2.9315\end{align*} is about \begin{align*}0.0262\end{align*}.

Example:

In the previous example, the \begin{align*}P\end{align*}-value for testing that the mean weight of cheese wasn’t \begin{align*}72\;\mathrm{ounces}\end{align*} was determined.

a) State the hypotheses.

b) Would the null hypothesis be rejected at the \begin{align*}10\%\end{align*} level? The \begin{align*}5\%\end{align*} level? The \begin{align*}1\%\end{align*} level?

Solution:

a) \begin{align*}H_0:\end{align*} The mean weight of cheese, \begin{align*}\mu\end{align*} is \begin{align*}72\;\mathrm{ounces}\end{align*}.

\begin{align*}H_\alpha: \mu \ne 72\end{align*}

b) Because the \begin{align*}P\end{align*}-value of \begin{align*}0.0262\end{align*} is less than both

Show Hide Details
Description
Tags:
Subjects:
Date Created:
Feb 23, 2012