11.4: F-Tests in One-Way ANOVA (2024)

  1. Last updated
  2. Save as PDF
  • Page ID
    516
    • 11.4: F-Tests in One-Way ANOVA (1)
    • Anonymous
    • LibreTexts

    \( \newcommand{\vecs}[1]{\overset { \scriptstyle \rightharpoonup} {\mathbf{#1}}}\)

    \( \newcommand{\vecd}[1]{\overset{-\!-\!\rightharpoonup}{\vphantom{a}\smash{#1}}} \)

    \( \newcommand{\id}{\mathrm{id}}\) \( \newcommand{\Span}{\mathrm{span}}\)

    ( \newcommand{\kernel}{\mathrm{null}\,}\) \( \newcommand{\range}{\mathrm{range}\,}\)

    \( \newcommand{\RealPart}{\mathrm{Re}}\) \( \newcommand{\ImaginaryPart}{\mathrm{Im}}\)

    \( \newcommand{\Argument}{\mathrm{Arg}}\) \( \newcommand{\norm}[1]{\| #1 \|}\)

    \( \newcommand{\inner}[2]{\langle #1, #2 \rangle}\)

    \( \newcommand{\Span}{\mathrm{span}}\)

    \( \newcommand{\id}{\mathrm{id}}\)

    \( \newcommand{\Span}{\mathrm{span}}\)

    \( \newcommand{\kernel}{\mathrm{null}\,}\)

    \( \newcommand{\range}{\mathrm{range}\,}\)

    \( \newcommand{\RealPart}{\mathrm{Re}}\)

    \( \newcommand{\ImaginaryPart}{\mathrm{Im}}\)

    \( \newcommand{\Argument}{\mathrm{Arg}}\)

    \( \newcommand{\norm}[1]{\| #1 \|}\)

    \( \newcommand{\inner}[2]{\langle #1, #2 \rangle}\)

    \( \newcommand{\Span}{\mathrm{span}}\) \( \newcommand{\AA}{\unicode[.8,0]{x212B}}\)

    \( \newcommand{\vectorA}[1]{\vec{#1}} % arrow\)

    \( \newcommand{\vectorAt}[1]{\vec{\text{#1}}} % arrow\)

    \( \newcommand{\vectorB}[1]{\overset { \scriptstyle \rightharpoonup} {\mathbf{#1}}}\)

    \( \newcommand{\vectorC}[1]{\textbf{#1}}\)

    \( \newcommand{\vectorD}[1]{\overrightarrow{#1}}\)

    \( \newcommand{\vectorDt}[1]{\overrightarrow{\text{#1}}}\)

    \( \newcommand{\vectE}[1]{\overset{-\!-\!\rightharpoonup}{\vphantom{a}\smash{\mathbf {#1}}}} \)

    \( \newcommand{\vecs}[1]{\overset { \scriptstyle \rightharpoonup} {\mathbf{#1}}}\)

    \( \newcommand{\vecd}[1]{\overset{-\!-\!\rightharpoonup}{\vphantom{a}\smash{#1}}} \)

    \(\newcommand{\avec}{\mathbf a}\) \(\newcommand{\bvec}{\mathbf b}\) \(\newcommand{\cvec}{\mathbf c}\) \(\newcommand{\dvec}{\mathbf d}\) \(\newcommand{\dtil}{\widetilde{\mathbf d}}\) \(\newcommand{\evec}{\mathbf e}\) \(\newcommand{\fvec}{\mathbf f}\) \(\newcommand{\nvec}{\mathbf n}\) \(\newcommand{\pvec}{\mathbf p}\) \(\newcommand{\qvec}{\mathbf q}\) \(\newcommand{\svec}{\mathbf s}\) \(\newcommand{\tvec}{\mathbf t}\) \(\newcommand{\uvec}{\mathbf u}\) \(\newcommand{\vvec}{\mathbf v}\) \(\newcommand{\wvec}{\mathbf w}\) \(\newcommand{\xvec}{\mathbf x}\) \(\newcommand{\yvec}{\mathbf y}\) \(\newcommand{\zvec}{\mathbf z}\) \(\newcommand{\rvec}{\mathbf r}\) \(\newcommand{\mvec}{\mathbf m}\) \(\newcommand{\zerovec}{\mathbf 0}\) \(\newcommand{\onevec}{\mathbf 1}\) \(\newcommand{\real}{\mathbb R}\) \(\newcommand{\twovec}[2]{\left[\begin{array}{r}#1 \\ #2 \end{array}\right]}\) \(\newcommand{\ctwovec}[2]{\left[\begin{array}{c}#1 \\ #2 \end{array}\right]}\) \(\newcommand{\threevec}[3]{\left[\begin{array}{r}#1 \\ #2 \\ #3 \end{array}\right]}\) \(\newcommand{\cthreevec}[3]{\left[\begin{array}{c}#1 \\ #2 \\ #3 \end{array}\right]}\) \(\newcommand{\fourvec}[4]{\left[\begin{array}{r}#1 \\ #2 \\ #3 \\ #4 \end{array}\right]}\) \(\newcommand{\cfourvec}[4]{\left[\begin{array}{c}#1 \\ #2 \\ #3 \\ #4 \end{array}\right]}\) \(\newcommand{\fivevec}[5]{\left[\begin{array}{r}#1 \\ #2 \\ #3 \\ #4 \\ #5 \\ \end{array}\right]}\) \(\newcommand{\cfivevec}[5]{\left[\begin{array}{c}#1 \\ #2 \\ #3 \\ #4 \\ #5 \\ \end{array}\right]}\) \(\newcommand{\mattwo}[4]{\left[\begin{array}{rr}#1 \amp #2 \\ #3 \amp #4 \\ \end{array}\right]}\) \(\newcommand{\laspan}[1]{\text{Span}\{#1\}}\) \(\newcommand{\bcal}{\cal B}\) \(\newcommand{\ccal}{\cal C}\) \(\newcommand{\scal}{\cal S}\) \(\newcommand{\wcal}{\cal W}\) \(\newcommand{\ecal}{\cal E}\) \(\newcommand{\coords}[2]{\left\{#1\right\}_{#2}}\) \(\newcommand{\gray}[1]{\color{gray}{#1}}\) \(\newcommand{\lgray}[1]{\color{lightgray}{#1}}\) \(\newcommand{\rank}{\operatorname{rank}}\) \(\newcommand{\row}{\text{Row}}\) \(\newcommand{\col}{\text{Col}}\) \(\renewcommand{\row}{\text{Row}}\) \(\newcommand{\nul}{\text{Nul}}\) \(\newcommand{\var}{\text{Var}}\) \(\newcommand{\corr}{\text{corr}}\) \(\newcommand{\len}[1]{\left|#1\right|}\) \(\newcommand{\bbar}{\overline{\bvec}}\) \(\newcommand{\bhat}{\widehat{\bvec}}\) \(\newcommand{\bperp}{\bvec^\perp}\) \(\newcommand{\xhat}{\widehat{\xvec}}\) \(\newcommand{\vhat}{\widehat{\vvec}}\) \(\newcommand{\uhat}{\widehat{\uvec}}\) \(\newcommand{\what}{\widehat{\wvec}}\) \(\newcommand{\Sighat}{\widehat{\Sigma}}\) \(\newcommand{\lt}{<}\) \(\newcommand{\gt}{>}\) \(\newcommand{\amp}{&}\) \(\definecolor{fillinmathshade}{gray}{0.9}\)

    Learning Objectives
    • To understand how to use an \(F\)-test to judge whether several population means are all equal

    In Chapter 9, we saw how to compare two population means \(\mu _1\) and \(\mu _2\).

    In this section we will learn to compare three or more population means at the same time, which is often of interest in practical applications. For example, an administrator at a university may be interested in knowing whether student grade point averages are the same for different majors. In another example, an oncologist may be interested in knowing whether patients with the same type of cancer have the same average survival times under several different competing cancer treatments.

    In general, suppose there are \(K\) normal populations with possibly different means, \(μ_1 , μ_2 , \ldots, μ_K\), but all with the same variance \(σ^2\). The study question is whether all the \(K\) population means are the same. We formulate this question as the test of hypotheses

    \[H_0: \mu _1=\mu _2=\cdots =\mu _K\\ vs.\\ H_a: \text{not all K population means are equal} \nonumber \]

    To perform the test \(K\) independent random samples are taken from the \(K\) normal populations. The \(K\) sample means, the \(K\) sample variances, and the \(K\) sample sizes are summarized in the table:

    Population Sample Size Sample Mean Sample Variance
    \(1\) \(n_1\) \(\bar{x_1}\) \(s_{1}^{2}\)
    \(2\) \(n_2\) \(\bar{x_2}\) \(s_{2}^{2}\)
    \(\vdots \) \(\vdots \) \(\vdots \) \(\vdots \)
    \(K\) \(n_K\) \(\bar{x_K}\) \(s_{K}^{2}\)

    Define the following quantities:

    Definitions

    The combined sample size:

    \[n=n_1+n_2+ \ldots + n_K \nonumber \]

    The mean of the combined sample of all \(n\) observations:

    \[x= \dfrac{\displaystyle Σx}{n} = \dfrac{n_1 \overline{x} + n_2 \overline{x}_2 + \ldots + n_K \overline{x}_K}{n} \nonumber \]

    The mean square for treatment:

    \[MST=\dfrac{n_1(\overline{x}_1−\overline{x})^2 + n_2(\overline{x}_2−\overline{x})^2 + \ldots  + n_K (\overline{x}K−\overline{x})^2}{K−1} \nonumber \]

    The mean square for error:

    \[MSE= \dfrac{(n_1−1)s^2_1 + (n_2−1)s^2_2 + \ldots + (n_{K−1})s^2_K}{n−K} \nonumber \]

    \(MST\) can be thought of as the variance between the \(K\) individual independent random samples and \(MSE\) as the variance within the samples. This is the reason for the name “analysis of variance,” universally abbreviated ANOVA. The adjective “one-way” has to do with the fact that the sampling scheme is the simplest possible, that of taking one random sample from each population under consideration. If the means of the \(K\) populations are all the same then the two quantities \(MST\) and \(MSE\) should be close to the same, so the null hypothesis will be rejected if the ratio of these two quantities is significantly greater than \(1\). This yields the following test statistic and methods and conditions for its use.

    Test Statistic for Testing the Null Hypothesis that \(K\) Population Means Are Equal

    \[F =\dfrac{MST}{MSE} \nonumber \]

    If the \(K\) populations are normally distributed with a common variance and if \(H_0: \mu _1=\mu _2=\cdots =\mu _K\) is true then under independent random sampling \(F\) approximately follows an \(F\)-distribution with degrees of freedom \(df_1=K-1\) and \(df_2=n-K\).

    The test is right-tailed: \(H_0\) is rejected at level of significance α if \(F≥F_α\).

    As always the test is performed using the usual five-step procedure.

    Example \(\PageIndex{1}\)

    The average of grade point averages (GPAs) of college courses in a specific major is a measure of difficulty of the major. An educator wishes to conduct a study to find out whether the difficulty levels of different majors are the same. For such a study, a random sample of major grade point averages (GPA) of \(11\) graduating seniors at a large university is selected for each of the four majors mathematics, English, education, and biology. The data are given in Table \(\PageIndex{1}\). Test, at the \(5\%\) level of significance, whether the data contain sufficient evidence to conclude that there are differences among the average major GPAs of these four majors.

    Table \(\PageIndex{1}\): Difficulty Levels of College Majors
    Mathematics English Education Biology
    2.59 3.64 4.00 2.78
    3.13 3.19 3.59 3.51
    2.97 3.15 2.80 2.65
    2.50 3.78 2.39 3.16
    2.53 3.03 3.47 2.94
    3.29 2.61 3.59 2.32
    2.53 3.20 3.74 2.58
    3.17 3.30 3.77 3.21
    2.70 3.54 3.13 3.23
    3.88 3.25 3.00 3.57
    2.64 4.00 3.47 3.22
    Solution
    • Step 1. The test of hypotheses is \[H_0: \mu _1=\mu _2=\mu _3=\mu _4\\ vs.\\ H_a: \text{not all four population means are equal}\; @\; \alpha =0.05 \nonumber \]
    • Step 2. The test statistic is \(F=MST/MSE\) with (since \(n=44\) and \(K=4\)) degrees of freedom \(df_1=K-1=4-1=3\) and \(df_2=n-K=44-4=40\).
    • Step 3. If we index the population of mathematics majors by \(1\), English majors by \(2\), education majors by \(3\), and biology majors by \(4\), then the sample sizes, sample means, and sample variances of the four samples in Table \(\PageIndex{1}\) are summarized (after rounding for simplicity) by:
    Major Sample Size Sample Mean Sample Variance
    Mathematics \(n_1=11\) \(\bar{x_1}=2.90\) \(s_{1}^{2}=0.188\)
    English \(n_2=11\) \(\bar{x_2}=3.34\) \(s_{2}^{2}=0.148\)
    Education \(n_3=11\) \(\bar{x_3}=3.36\) \(s_{3}^{2}=0.229\)
    Biology \(n_4=11\) \(\bar{x_4}=3.02\) \(s_{4}^{2}=0.157\)

    The average of all \(44\) observations is (after rounding for simplicity) \(\overline{x}=3.15\). We compute (rounding for simplicity)

    \[\begin{align} MST &= \dfrac{11(2.90−3.15)^2+11(3.34−3.15)^2+11(3.36−3.15)^2+11(3.02−3.15)^2}{4−1} \nonumber \\[6pt] &=\dfrac{1.7556}{3} \nonumber \\[6pt] &=0.585 \nonumber \end{align} \nonumber \]

    and

    \[\begin{align} MSE &= \dfrac{(11−1)(0.188)+(11−1)(0.148)+(11−1)(0.229)+(11−1)(0.157)}{44−4} \nonumber \\[6pt] &=\dfrac{7.22}{40} \nonumber \\[6pt] &=0.181 \nonumber \end{align} \nonumber \]

    so that

    \[F=\dfrac{MST}{MSE}=\dfrac{0.585}{0.181}=3.232 \nonumber \]

    • Step 4. The test is right-tailed. The single critical value is (since \(df_1=3\) and \(df_2=40\)) \(F_\alpha =F_{0.05}=2.84\). Thus the rejection region is \([2.84,\infty )\), as illustrated in Figure \(\PageIndex{1}\).
    11.4: F-Tests in One-Way ANOVA (2)
    • Step 5. Since \(F=3.232>2.84\), we reject \(H_0\). The data provide sufficient evidence, at the \(5\%\) level of significance, to conclude that the averages of major GPAs for the four majors considered are not all equal.
    Example \(\PageIndex{2}\): Mice Survival Times

    A research laboratory developed two treatments which are believed to have the potential of prolonging the survival times of patients with an acute form of thymic leukemia. To evaluate the potential treatment effects \(33\) laboratory mice with thymic leukemia were randomly divided into three groups. One group received Treatment \(1\), one received Treatment \(2\), and the third was observed as a control group. The survival times of these mice are given in Table \(\PageIndex{2}\). Test, at the \(1\%\) level of significance, whether these data provide sufficient evidence to confirm the belief that at least one of the two treatments affects the average survival time of mice with thymic leukemia.

    Table \(\PageIndex{2}\) Mice Survival Times in Days
    Treatment \(1\) Treatment \(2\) Control
    71 75 77 81
    72 73 67 79
    75 72 79 73
    80 65 78 71
    60 63 81 75
    65 69 72 84
    63 64 71 77
    78 71 84 67
    91
    Solution
    • Step 1. The test of hypotheses is \[H_0: \mu _1=\mu _2=\mu _3\\ vs.\\ H_a: \text{not all three population means are equal}\; @\; \alpha =0.01 \nonumber \]
    • Step 2. The test statistic is \(F=\dfrac{MST}{MSE}\) with (since \(n=33\) and \(K=3\)) degrees of freedom \(df_1=K-1=3-1=2\) and \(df_2=n-K=33-3=30\).
    • Step 3. If we index the population of mice receiving Treatment \(1\) by \(1\), Treatment \(2\) by \(2\), and no treatment by \(3\), then the sample sizes, sample means, and sample variances of the three samples in Table \(\PageIndex{2}\) are summarized (after rounding for simplicity) by:
    Table \(\PageIndex{2}\): Mice Survival Times in Days
    Group Sample Size Sample Mean Sample Variance
    Treatment \(1\) \(n_1=16\) \(\bar{x_1}=69.75\) \(s_{1}^{2}=34.47\)
    Treatment \(2\) \(n_2=9\) \(\bar{x_2}=77.78\) \(s_{2}^{2}=52.69\)
    Control \(n_3=8\) \(\bar{x_3}=75.88\) \(s_{3}^{2}=30.69\)

    The average of all \(33\) observations is (after rounding for simplicity) \(\overline{x}=73.42\). We compute (rounding for simplicity)

    \[\begin{align*} MST &= \frac{16(69.75-73.42)^2+9(77.78-73.42)^2+8(75.88-73.42)^2}{31}\\ &= \frac{434.63}{2}\\ &= 217.50 \end{align*} \nonumber \]

    and

    \[\begin{align*} MSE &= \frac{(16-1)(34.47)+(9-1)(52.69)+(8-1)(30.69)}{33-3}\\ &= \frac{1153.4}{30}\\ &= 38.45\end{align*} \nonumber \]

    so that

    \[F=\dfrac{MST}{MSE}=\dfrac{217.50}{38.45}=5.65 \nonumber \]

    • Step 4. The test is right-tailed. The single critical value is \(F_\alpha =F_{0.01}=5.39\). Thus the rejection region is \([5.39,\infty )\), as illustrated in Figure \(\PageIndex{2}\).
    11.4: F-Tests in One-Way ANOVA (3)
    • Step 5. Since \(F=5.65>5.39\), we reject \(H_0\). The data provide sufficient evidence, at the \(1\%\) level of significance, to conclude that a treatment effect exists at least for one of the two treatments in increasing the mean survival time of mice with thymic leukemia.

    It is important to to note that, if the null hypothesis of equal population means is rejected, the statistical implication is that not all population means are equal. It does not however tell which population mean is different from which. The inference about where the suggested difference lies is most frequently made by a follow-up study.

    Key Takeaway
    • An \(F\)-test can be used to evaluate the hypothesis that the means of several normal populations, all with the same standard deviation, are identical.
    11.4: F-Tests in One-Way ANOVA (2024)

    FAQs

    What is a good F value in ANOVA? ›

    If the null hypothesis is true, you expect F to have a value close to 1.0 most of the time. A large F ratio means that the variation among group means is more than you'd expect to see by chance.

    What if the F value is greater than 1 in ANOVA? ›

    If the F-score is much greater than one, the variance between is probably the source of most of the variance in the total sample, and the samples probably come from populations with different means.

    What is the F-test statistic in a one-way ANOVA? ›

    The F-test (for ANOVA) is a statistical test for testing the equality of k population means. The one-way ANOVA F-test is a statistical test for testing the equality of k population means from 3 or more groups within one variable or factor.

    What is the acceptable F test value? ›

    A general rule of thumb that is often used in regression analysis is that if F > 2.5 then we can reject the null hypothesis.

    How to know if ANOVA is significant? ›

    ANOVA results are considered significant if they show a significant difference between groups or conditions being compared. ANOVA results are considered significant if the p-value is less than the chosen significance level (usually 0.05).

    What does the F statistic in a one-way ANOVA represent? ›

    The test statistic for ANOVA is an F statistic. It is essentially a ratio of between-group variation to within-group variation. ANOVA tells you whether the mean of at least one group is significantly different from those of the other groups, but it does not tell you which mean.

    How to interpret F value and p-value? ›

    A big F, with a small p-value, means that the null hypothesis is discredited, and we would assert that the means are significantly different (while a small F, with a big p-value indicates that they are not significantly different).

    What is the F ratio in a one-way ANOVA? ›

    The F-ratio is the ratio of the between group variance to the within group variance. The F-ratio is used in an ANOVA (Analysis of Variance) that provides more insight into data compared to using only the mean or median.

    What is T value and F value in ANOVA? ›

    In student's t test, calculated t value is ratio of mean difference and standard error, whereas in the ANOVA test, calculated F value is ratio of the variability between groups with the variability of the observations within the groups.[1,4]

    How many F tests are in a two-way ANOVA? ›

    The correct answer is b. 2. Unlike a one-way ANOVA, a two-way ANOVA presents 2 F statistics. This test statistic summarizes whether there is a different effect from any of the levels of any of the factors.

    What is a high F value in ANOVA? ›

    The F-value in an ANOVA is calculated as: variation between sample means / variation within the samples. The higher the F-value in an ANOVA, the higher the variation between sample means relative to the variation within the samples. The higher the F-value, the lower the corresponding p-value.

    How to calculate F-test value? ›

    The F-test is a type of hypothesis testing that uses the F-statistic to analyze data variance in two samples or populations. The F-statistic, or F-value, is calculated as follows: F = σ 1 σ 2 , or Variance 1/Variance 2. Hypothesis testing of variance relies directly upon the F-distribution data for its comparisons.

    What is the standard error of the ANOVA? ›

    The standard error of the estimate (se), also known as the root mean square error or the standard error of the regression, can be calculated from the ANOVA table. The se measures the distance between values predicted from the estimated regression and the observed values of the dependent variable.

    What if the F-value is close to 1? ›

    F-statistics are the ratio of two variances that are approximately the same value when the null hypothesis is true, which yields F-statistics near 1.

    What does the F ratio tell you in ANOVA? ›

    The F-ratio is the ratio of the between group variance to the within group variance. The F-ratio is used in an ANOVA (Analysis of Variance) that provides more insight into data compared to using only the mean or median.

    Can F-value be less than 1? ›

    When the error variance is higher than the effect variance, then we will always get an F-value less than one. You can see that we often got F-values less than one in the simulation. This is sensible, after all we were simulating samples coming from the very same distribution.

    What is the F effect size in ANOVA? ›

    Effect size for F-ratios in analysis of variance

    The effect size used in analysis of variance is defined by the ratio of population standard deviations. Although Cohen's f is defined as above it is usually computed by taking the square root of f2.

    References

    Top Articles
    Latest Posts
    Article information

    Author: Lilliana Bartoletti

    Last Updated:

    Views: 6148

    Rating: 4.2 / 5 (53 voted)

    Reviews: 84% of readers found this page helpful

    Author information

    Name: Lilliana Bartoletti

    Birthday: 1999-11-18

    Address: 58866 Tricia Spurs, North Melvinberg, HI 91346-3774

    Phone: +50616620367928

    Job: Real-Estate Liaison

    Hobby: Graffiti, Astronomy, Handball, Magic, Origami, Fashion, Foreign language learning

    Introduction: My name is Lilliana Bartoletti, I am a adventurous, pleasant, shiny, beautiful, handsome, zealous, tasty person who loves writing and wants to share my knowledge and understanding with you.