# 11-3. F-tests for Equality of Two Variances

## 1. *F*-Distributions

Another important and useful family of distributions in statistics is the family of *F*-distributions. Each member of the *F*-distribution family is specified by a pair of parameters called *degrees of freedom* and denoted $$df\_1$$ and $$df\_2$$ . Figure 11.7 "Many " shows several *F*-distributions for different pairs of degrees of freedom. An **F random variable** is a random variable that assumes only positive values and follows an *F*-distribution.

Figure 11.7 Many *F*-Distributions

![](https://saylordotorg.github.io/text_introductory-statistics/section_15/75fd6a2d869b9403de8408b42a054db9.jpg)

The parameter $$df\_1$$ is often referred to as the *numerator* degrees of freedom and the parameter $$df\_2$$ as the *denominator* degrees of freedom. It is important to keep in mind that they are not interchangeable. For example, the *F*-distribution with degrees of freedom $$df\_1=3$$ and $$df\_2=8$$ is a different distribution from the *F*-distribution with degrees of freedom $$df\_1=8$$ and $$df\_2=3$$.

####

> *The value of the* *F random variable* *F* *with degrees of freedom* $$df\_1$$ *and* $$df\_2$$ *that cuts off a right tail of area* *c* *is denoted* $$F\_c$$ *and is called a* **critical value**. *See Figure 11.8.*
>
> Figure 11.8 $$F\_c$$ Illustrated

![](https://saylordotorg.github.io/text_introductory-statistics/section_15/59a41857cbaae56b820472577df06b26.jpg)

Tables containing the values of $$F\_c$$ are given in Chapter 11 "Chi-Square Tests and ". Each of the tables is for a fixed collection of values of *c*, either 0.900, 0.950, 0.975, 0.990, and 0.995 (yielding what are called “lower” critical values), or 0.005, 0.010, 0.025, 0.050, and 0.100 (yielding what are called “upper” critical values). In each table critical values are given for various pairs $$(df\_1,df\_2)$$ . We illustrate the use of the tables with several examples.

### [F-Distribution Tables](http://www.socr.ucla.edu/Applets.dir/F_Table.html#FTable0.05)

**EXAMPLE 3.**  Suppose *F* is an *F* random variable with degrees of freedom $$df\_1=5$$ and $$df\_2=4$$ . Use the tables to find

1. $$F\_{0.10}$$&#x20;
2. $$F\_{0.95}$$&#x20;

**\[ Solution ]**

1. The column headings of all the tables contain $$df\_1=5$$ . Look for the table for which 0.10 is one of the entries on the extreme left (a table of upper critical values) and that has a row heading $$df\_2=4$$ in the left margin of the table. A portion of the relevant table is provided. The entry in the intersection of the column with heading $$df\_1=5$$ and the row with the headings 0.10 and $$df\_2=4$$, which is shaded in the table provided, is the answer, $$F\_{0.10}=4.05$$ .

   | *F* Tail Area | $$df\_1$$ | 1              | 2              |  ⋅ ⋅ ⋅  · · ·  | 5        |  ⋅ ⋅ ⋅  · · ·  |
   | ------------- | --------- | -------------- | -------------- | -------------- | -------- | -------------- |
   | $$df\_2$$     |           |                |                |                |          |                |
   | ⋮             | ⋮         | ⋮              | ⋮              | ⋮              | ⋮        | ⋮              |
   | 0.005         | 4         |  ⋅ ⋅ ⋅  · · ·  |  ⋅ ⋅ ⋅  · · ·  |  ⋅ ⋅ ⋅  · · ·  | 22.5     |  ⋅ ⋅ ⋅  · · ·  |
   | 0.01          | 4         |  ⋅ ⋅ ⋅  · · ·  |  ⋅ ⋅ ⋅  · · ·  |  ⋅ ⋅ ⋅  · · ·  | 15.5     |  ⋅ ⋅ ⋅  · · ·  |
   | 0.025         | 4         |  ⋅ ⋅ ⋅  · · ·  |  ⋅ ⋅ ⋅  · · ·  |  ⋅ ⋅ ⋅  · · ·  | 9.36     |  ⋅ ⋅ ⋅  · · ·  |
   | 0.05          | 4         |  ⋅ ⋅ ⋅  · · ·  |  ⋅ ⋅ ⋅  · · ·  |  ⋅ ⋅ ⋅  · · ·  | 6.26     |  ⋅ ⋅ ⋅  · · ·  |
   | 0.10          | 4         |  ⋅ ⋅ ⋅  · · ·  |  ⋅ ⋅ ⋅  · · ·  |  ⋅ ⋅ ⋅  · · ·  | $$4.05$$ |  ⋅ ⋅ ⋅  · · ·  |
   | ⋮             | ⋮         | ⋮              | ⋮              | ⋮              | ⋮        | ⋮              |
2. Look for the table for which 0.95 is one of the entries on the extreme left (a table of lower critical values) and that has a row heading $$df\_2=4$$ in the left margin of the table. A portion of the relevant table is provided. The entry in the intersection of the column with heading $$df\_1=5$$ and the row with the headings 0.95 and $$df\_2=4$$, which is shaded in the table provided, is the answer, $$F\_{0.95}=0.19$$ .

   | *F* Tail Area | $$df\_1$$ | 1              | 2              |  ⋅ ⋅ ⋅  · · ·  | 5        |  ⋅ ⋅ ⋅  · · ·  |
   | ------------- | --------- | -------------- | -------------- | -------------- | -------- | -------------- |
   | $$df\_2$$     |           |                |                |                |          |                |
   | ⋮             | ⋮         | ⋮              | ⋮              | ⋮              | ⋮        | ⋮              |
   | 0.90          | 4         |  ⋅ ⋅ ⋅  · · ·  |  ⋅ ⋅ ⋅  · · ·  |  ⋅ ⋅ ⋅  · · ·  | 0.28     |  ⋅ ⋅ ⋅  · · ·  |
   | 0.95          | 4         |  ⋅ ⋅ ⋅  · · ·  |  ⋅ ⋅ ⋅  · · ·  |  ⋅ ⋅ ⋅  · · ·  | $$0.19$$ |  ⋅ ⋅ ⋅  · · ·  |
   | 0.975         | 4         |  ⋅ ⋅ ⋅  · · ·  |  ⋅ ⋅ ⋅  · · ·  |  ⋅ ⋅ ⋅  · · ·  | 0.14     |  ⋅ ⋅ ⋅  · · ·  |
   | 0.99          | 4         |  ⋅ ⋅ ⋅  · · ·  |  ⋅ ⋅ ⋅  · · ·  |  ⋅ ⋅ ⋅  · · ·  | 0.09     |  ⋅ ⋅ ⋅  · · ·  |
   | 0.995         | 4         |  ⋅ ⋅ ⋅  · · ·  |  ⋅ ⋅ ⋅  · · ·  |  ⋅ ⋅ ⋅  · · ·  | 0.06     |  ⋅ ⋅ ⋅  · · ·  |
   | ⋮             | ⋮         | ⋮              | ⋮              | ⋮              | ⋮        | ⋮              |

{% tabs %}
{% tab title="R Source" %}

```
library(Rstat)

df_1 <- 5
df_2 <- 4

pv <- c(0.90, 0.05)
f.quant(df_1, df_2, pv)
```

{% endtab %}

{% tab title="Result" %}

```
> f.quant(df_1, df_2, pv)
##   0.9   0.05 
## 4.0506 0.1926 
```

![](https://2234305379-files.gitbook.io/~/files/v0/b/gitbook-legacy-files/o/assets%2F-Lqmt1Wp54aKlZ5eVcaS%2F-Lrm97gd4pgBMhlgKIX4%2F-LrmAzcf-vCYv3SbLNxf%2Fimage.png?alt=media\&token=b05a7dde-7f65-48e6-9ecc-f83cc17b208e)
{% endtab %}
{% endtabs %}

**EXAMPLE 4.**   Suppose $$F$$ is an *F* random variable with degrees of freedom $$df\_1=2$$ and $$df\_2=20$$. Let $$α=0.05$$ . Use the tables to find

1. $$F\_α$$&#x20;
2. $$F\_{α∕2}$$&#x20;
3. $$F\_{1−α}$$&#x20;
4. $$F\_{1−α∕2}$$&#x20;

**\[ Solution ]**

1. The column headings of all the tables contain $$df\_1=2$$ . Look for the table for which $$α=0.05$$ is one of the entries on the extreme left (a table of upper critical values) and that has a row heading $$df\_2=20$$ in the left margin of the table. A portion of the relevant table is provided. The shaded entry, in the intersection of the column with heading $$df\_1=2$$ and the row with the headings 0.05 and $$df\_2=20$$ is the answer, $$F\_{0.05}=3.49$$ .

   | *F* Tail Area | $$df\_1$$ | 1              | 2        |  ⋅ ⋅ ⋅  · · ·  |
   | ------------- | --------- | -------------- | -------- | -------------- |
   | $$df\_2$$     |           |                |          |                |
   | ⋮             | ⋮         | ⋮              | ⋮        | ⋮              |
   | 0.005         | 20        |  ⋅ ⋅ ⋅  · · ·  | 6.99     |  ⋅ ⋅ ⋅  · · ·  |
   | 0.01          | 20        |  ⋅ ⋅ ⋅  · · ·  | 5.85     |  ⋅ ⋅ ⋅  · · ·  |
   | 0.025         | 20        |  ⋅ ⋅ ⋅  · · ·  | 4.46     |  ⋅ ⋅ ⋅  · · ·  |
   | 0.05          | 20        |  ⋅ ⋅ ⋅  · · ·  | $$3.49$$ |  ⋅ ⋅ ⋅  · · ·  |
   | 0.10          | 20        |  ⋅ ⋅ ⋅  · · ·  | 2.59     |  ⋅ ⋅ ⋅  · · ·  |
   | ⋮             | ⋮         | ⋮              | ⋮        | ⋮              |
2. Look for the table for which $$α∕2=0.025$$ is one of the entries on the extreme left (a table of upper critical values) and that has a row heading $$df\_2=20$$ in the left margin of the table. A portion of the relevant table is provided. The shaded entry, in the intersection of the column with heading $$df\_1=2$$ and the row with the headings 0.025 and $$df\_2=20$$ is the answer, $$F\_{0.025}=4.46$$ .

   | *F* Tail Area | $$df\_1$$ | 1              | 2        |  ⋅ ⋅ ⋅  · · ·  |
   | ------------- | --------- | -------------- | -------- | -------------- |
   | $$df\_2$$     |           |                |          |                |
   | ⋮             | ⋮         | ⋮              | ⋮        | ⋮              |
   | 0.005         | 20        |  ⋅ ⋅ ⋅  · · ·  | 6.99     |  ⋅ ⋅ ⋅  · · ·  |
   | 0.01          | 20        |  ⋅ ⋅ ⋅  · · ·  | 5.85     |  ⋅ ⋅ ⋅  · · ·  |
   | 0.025         | 20        |  ⋅ ⋅ ⋅  · · ·  | $$4.46$$ |  ⋅ ⋅ ⋅  · · ·  |
   | 0.05          | 20        |  ⋅ ⋅ ⋅  · · ·  | 3.49     |  ⋅ ⋅ ⋅  · · ·  |
   | 0.10          | 20        |  ⋅ ⋅ ⋅  · · ·  | 2.59     |  ⋅ ⋅ ⋅  · · ·  |
   | ⋮             | ⋮         | ⋮              | ⋮        | ⋮              |
3. Look for the table for which $$(1−α)=0.95$$ is one of the entries on the extreme left (a table of lower critical values) and that has a row heading $$df\_2=20$$ in the left margin of the table. A portion of the relevant table is provided. The shaded entry, in the intersection of the column with heading $$df\_1=2$$ and the row with the headings 0.95 and $$df\_2=20$$ is the answer, F0.95=0.05.F0.95=0.05.

   | *F* Tail Area | $$df\_1$$ | 1              | 2        |  ⋅ ⋅ ⋅  · · ·  |
   | ------------- | --------- | -------------- | -------- | -------------- |
   | $$df\_2$$     |           |                |          |                |
   | ⋮             | ⋮         | ⋮              | ⋮        | ⋮              |
   | 0.90          | 20        |  ⋅ ⋅ ⋅  · · ·  | 0.11     |  ⋅ ⋅ ⋅  · · ·  |
   | 0.95          | 20        |  ⋅ ⋅ ⋅  · · ·  | $$0.05$$ |  ⋅ ⋅ ⋅  · · ·  |
   | 0.975         | 20        |  ⋅ ⋅ ⋅  · · ·  | 0.03     |  ⋅ ⋅ ⋅  · · ·  |
   | 0.99          | 20        |  ⋅ ⋅ ⋅  · · ·  | 0.01     |  ⋅ ⋅ ⋅  · · ·  |
   | 0.995         | 20        |  ⋅ ⋅ ⋅  · · ·  | 0.01     |  ⋅ ⋅ ⋅  · · ·  |
   | ⋮             | ⋮         | ⋮              | ⋮        | ⋮              |
4. Look for the table for which $$(1−α∕2)=0.975$$ is one of the entries on the extreme left (a table of lower critical values) and that has a row heading $$df\_2=20$$ in the left margin of the table. A portion of the relevant table is provided. The shaded entry, in the intersection of the column with heading $$df\_1=2$$ and the row with the headings 0.975 and $$df\_2=20$$ is the answer, $$F\_{0.975}=0.03$$ .

   | *F* Tail Area | $$df\_1$$ | 1              | 2        |  ⋅ ⋅ ⋅  · · ·  |
   | ------------- | --------- | -------------- | -------- | -------------- |
   | $$df\_2$$     |           |                |          |                |
   | ⋮             | ⋮         | ⋮              | ⋮        | ⋮              |
   | 0.90          | 20        |  ⋅ ⋅ ⋅  · · ·  | 0.11     |  ⋅ ⋅ ⋅  · · ·  |
   | 0.95          | 20        |  ⋅ ⋅ ⋅  · · ·  | 0.05     |  ⋅ ⋅ ⋅  · · ·  |
   | 0.975         | 20        |  ⋅ ⋅ ⋅  · · ·  | $$0.03$$ |  ⋅ ⋅ ⋅  · · ·  |
   | 0.99          | 20        |  ⋅ ⋅ ⋅  · · ·  | 0.01     |  ⋅ ⋅ ⋅  · · ·  |
   | 0.995         | 20        |  ⋅ ⋅ ⋅  · · ·  | 0.01     |  ⋅ ⋅ ⋅  · · ·  |
   | ⋮             | ⋮         | ⋮              | ⋮        | ⋮              |

{% tabs %}
{% tab title="R Source" %}

```
library(Rstat)

df_1 <- 2
df_2 <- 20

pv <- c(0.95, 0.975, 0.05, 0.025)
f.quant(df_1, df_2, pv)

```

{% endtab %}

{% tab title="Result" %}

```
> f.quant(df_1, df_2, pv)
##   0.95  0.975   0.05  0.025 
## 3.4928 4.4613 0.0514 0.0253 
```

![](https://2234305379-files.gitbook.io/~/files/v0/b/gitbook-legacy-files/o/assets%2F-Lqmt1Wp54aKlZ5eVcaS%2F-Lrm97gd4pgBMhlgKIX4%2F-LrmBjSv71_wbtO7Aeqz%2Fimage.png?alt=media\&token=fdb630e8-b042-40a3-9c0a-48d9faded831)
{% endtab %}
{% endtabs %}

> A fact that sometimes allows us to find a critical value from a table that we could not read otherwise is:
>
> If $$F\_u(r,s)$$ denotes the value of the *F*-distribution with degrees of freedom $$df\_1=r$$ and $$df\_2=s$$&#x20;
>
> that cuts off a right tail of area *u*, then
>
> &#x20;                                                              $$F\_c(k,ℓ)= \frac {1}{F\_{1−c}(ℓ,k)}$$&#x20;

**EXAMPLE 5.**  Use the tables to find

1. $$F\_{0.01}$$ for an *F* random variable with $$df\_1=13$$ and $$df\_2=8$$&#x20;
2. $$F\_{0.975}$$ for an *F* random variable with $$df\_1=40$$ and $$df\_2=10$$&#x20;

**\[ Solution ]**

1. There is no table with $$df\_1=13$$, but there is one with $$df\_2=8$$. Thus we use the fact that\
   \
   &#x20;                             $$F\_{0.01}(13, 8)= \frac{1}{F\_{0.99}(8,13)}$$ <br>

Using the relevant table we find that $$F\_{0.99}(8,13)=0.18$$ , \
hence  $$F\_{0.01}(13,8)=0.18^{-1} = 5.556.$$<br>

{% tabs %}
{% tab title="R Source" %}

```
library(Rstat)

df_1 <- 13
df_2 <- 8
alpha <- 0.01

pv <- c(1-alpha)
f.quant(df_1, df_2, pv)

pv_1 <- c(alpha)
f.quant(df_2, df_1, pv_1)

```

{% endtab %}

{% tab title="Result" %}

```
> pv <- c(1-alpha)
> f.quant(df_1, df_2, pv)
##   0.99 
## 5.6089 
> pv_1 <- c(alpha)
> f.quant(df_2, df_1, pv_1) 
##   0.01 
## 0.1783 
```

![](https://2234305379-files.gitbook.io/~/files/v0/b/gitbook-legacy-files/o/assets%2F-Lqmt1Wp54aKlZ5eVcaS%2F-LrmDOE6pshB2YKbn8M_%2F-LrmDtVnFuRjC6pc-pbC%2Fimage.png?alt=media\&token=d05bdf69-9a87-4a62-b152-04700da4e8c7)

![](https://2234305379-files.gitbook.io/~/files/v0/b/gitbook-legacy-files/o/assets%2F-Lqmt1Wp54aKlZ5eVcaS%2F-LrmDOE6pshB2YKbn8M_%2F-LrmDy0P9DHXHdUZEsWW%2Fimage.png?alt=media\&token=a4d32b06-36e1-44d3-b1c0-d8569e4ddb24)
{% endtab %}
{% endtabs %}

2\. There is no table with $$df\_1=40$$, but there is one with $$df\_2=10$$. Thus we use the fact that\
\
&#x20;                               $$F\_{0.975}(40,10)= \frac{1}{F\_{0.025}(10,40)}$$

\
Using the relevant table we find that $$F\_{0.025}(10,40)=2.3882$$ , \
hence $$F\_{0.975}(10,40)=2.3882^{-1}=0.4187$$

{% tabs %}
{% tab title="R Source" %}

```
library(Rstat)

df_1 <- 40
df_2 <- 10
alpha <- 0.975

pv <- c(1-alpha)
f.quant(df_1, df_2, pv)

pv_1 <- c(alpha)
f.quant(df_2, df_1, pv_1)


```

{% endtab %}

{% tab title="Result" %}

```
> pv <- c(1-alpha)
> f.quant(df_1, df_2, pv)
##  0.025 
## 0.4187 
> pv_1 <- c(alpha)
> f.quant(df_2, df_1, pv_1)
##  0.975 
## 2.3882 
```

![](https://2234305379-files.gitbook.io/~/files/v0/b/gitbook-legacy-files/o/assets%2F-Lqmt1Wp54aKlZ5eVcaS%2F-LrmG5hL2FYLSaKb2HbT%2F-LrmGTBAh0qzDPzXGl3d%2Fimage.png?alt=media\&token=17f63309-68e0-487b-8984-556171640284)

![](https://2234305379-files.gitbook.io/~/files/v0/b/gitbook-legacy-files/o/assets%2F-Lqmt1Wp54aKlZ5eVcaS%2F-LrmG5hL2FYLSaKb2HbT%2F-LrmGYenEug129URYgKy%2Fimage.png?alt=media\&token=56f5b969-dde0-4f19-975d-c843323f076a)
{% endtab %}
{% endtabs %}

## 2. *F*-Tests for Equality of Two Variances

In Chapter 9 "Two-Sample Problems" we saw how to test hypotheses about the difference between two population means $$μ\_1$$ and $$μ\_2$$ . In some practical situations the difference between the population standard deviations $$σ\_1$$ and $$σ\_2$$ is also of interest. Standard deviation measures the variability of a random variable. For example, if the random variable measures the size of a machined part in a manufacturing process, the size of standard deviation is one indicator of product quality. A smaller standard deviation among items produced in the manufacturing process is desirable since it indicates consistency in product quality.

For theoretical reasons it is easier to compare the squares of the population standard deviations, the population variances $$σ\_1^2$$ and $$σ\_2^2$$. This is not a problem, since $$σ\_1=σ\_2$$ precisely when $$σ^2\_1=σ\_2^2$$ , $$σ\_1<σ\_2$$ precisely when $$σ\_1^2<σ\_2^2$$ , and $$σ\_1>σ\_2$$ precisely when $$σ\_1^2>σ\_2^2$$ .

The null hypothesis always has the form $$H\_0:σ\_1=σ\_2$$. The three forms of the alternative hypothesis, with the terminology for each case, are:

| Form of $$H\_a$$           | Terminology  |
| -------------------------- | ------------ |
| $$H\_a:σ\_1^2>σ\_2^2$$     | Right-tailed |
| $$H\_a:σ\_1^2<σ\_2^2$$     | Left-tailed  |
| $$H\_a:σ\_1^2 \ne σ\_2^2$$ | Two-tailed   |

Just as when we test hypotheses concerning two population means, we take a random sample from each population, of sizes $$n\_1$$ and $$n\_2$$ , and compute the sample standard deviations $$s\_1$$ and $$s\_2$$ . In this context the samples are always independent. The populations themselves must be normally distributed.

> **Test Statistic for Hypothesis Tests Concerning the Difference Between Two Population Variances**
>
> &#x20;                                                           $$F= \frac{s^2\_1}{s\_2^2}$$&#x20;
>
> If the two populations are normally distributed and if  $$H\_0:σ^2\_1=σ\_2^2$$  is true then under independent sampling *F* approximately follows an *F*-distribution with degrees of freedom $$df\_1=(n\_1−1)$$ and $$df\_2=(n\_2−1)$$ .

A test based on the test statistic $$F$$ is called an *F*-test.

A most important point is that while the rejection region for a right-tailed test is exactly as in every other situation that we have encountered, because of the asymmetry in the *F*-distribution the critical value for a left-tailed test and the lower critical value for a two-tailed test have the special forms shown in the following table:

| Terminology  | Alternative Hypothesis     | Rejection Region                             |
| ------------ | -------------------------- | -------------------------------------------- |
| Right-tailed | $$H\_a:σ\_1^2>σ\_2^2$$     | $$F≥F\_α$$                                   |
| Left-tailed  | $$H\_a:σ\_1^2<σ\_2^2$$     | $$F≤F\_{1−α}$$                               |
| Two-tailed   | $$H\_a:σ\_1^2 \ne σ\_2^2$$ | $$F≤F\_{1−α∕2} \space or \space F≥F\_{α∕2}$$ |

Figure 11.9 "Rejection Regions: (a) Right-Tailed; (b) Left-Tailed; (c) Two-Tailed" illustrates these **rejection regions**.

Figure 11.9  Rejection Regions: (a) Right-Tailed; (b) Left-Tailed; (c) Two-Tailed

![](https://saylordotorg.github.io/text_introductory-statistics/section_15/da58e4f2977f4b9644c6f0c320af4429.jpg)

The test is performed using the usual five-step procedure described at the end of Section 8.1 "The Elements of Hypothesis Testing" in Chapter 8 "Testing Hypotheses".

**EXAMPLE 6.**  One of the quality measures of blood glucose meter strips is the consistency of the test results on the same sample of blood. The consistency is measured by the variance of the readings in repeated testing. Suppose two types of strips, *A* and *B*, are compared for their respective consistencies. We arbitrarily label the population of Type *A* strips Population 1 and the population of Type *B* strips Population 2. Suppose 15 Type *A* strips were tested with blood drops from a well-shaken vial and 20 Type *B* strips were tested with the blood from the same vial. The results are summarized in Table 11.16 "Two Types of Test Strips". Assume the glucose readings using Type *A* strips follow a normal distribution with variance σ21σ12 and those using Type *B* strips follow a normal distribution with variance with σ22.σ22. Test, at the 10% level of significance, whether the data provide sufficient evidence to conclude that the consistencies of the two types of strips are different.

TABLE 11.16 TWO TYPES OF TEST STRIPS

| Strip Type | Sample Size | Sample Variance |
| ---------- | ----------- | --------------- |
| *A*        | $$n\_1=16$$ | $$s\_1^2=2.09$$ |
| *B*        | $$n\_2=21$$ | $$s\_2^2=1.10$$ |

**\[ Solution ]**

* **Step 1.** The test of **hypotheses** is\
  \
  &#x20;                    $$H\_0:σ\_1=σ\_2$$\
  &#x20;              vs. $$H\_a:σ\_1^2 \ne σ\_2^2   \space \space @ α=0.10$$<br>
* **Step 2.** The distribution is the *F*-distribution with **degrees of freedom**

  &#x20;             $$df\_1=(16−1)=15$$ and $$df\_2=(21−1)=20$$ .<br>
* **Step 3.** The test is **two-tailed**. The **left or lower critical value** is\
  &#x20;                               $$F\_{1−α∕2}=F\_{0.95}=0.43$$ .\
  &#x20;            The right or upper critical value is \
  &#x20;                                $$F\_{α∕2}=F\_{0.05}=2.20$$ . \
  &#x20;            Thus the **rejection region** is $$\[0,−0.43]∪\[2.20,∞)$$ , \
  &#x20;            as illustrated in Figure 11.10 "Rejection Region and Test Statistic for ".

Figure 11.10  Rejection Region and Test Statistic for Note 11.27 "Example 6"

![](https://saylordotorg.github.io/text_introductory-statistics/section_15/5529f95c575a61746d7c6db537493c0f.jpg)

* **Step 4.** The value of the test statistic is\
  &#x20;               $$F=\frac{s^2\_1}{s^2\_2}=\frac{2.09}{1.10}=1.90$$ <br>
* **Step 5.** As shown in Figure 11.10 "Rejection Region and Test Statistic for ", \
  &#x20;             the test statistic 1.90 does not lie in the rejection region, so the decision is **not to reject** $$H\_0$$ .\
  &#x20;
* The data do not provide sufficient evidence, at the 10% level of significance, to **conclude** that there is a difference in the consistency, as measured by the variance, of the two types of test strips.

**EXAMPLE 7.**  In the context of Note 11.27 "Example 6", suppose Type *A* test strips are the current market leader and Type *B* test strips are a newly improved version of Type *A*. Test, at the 10% level of significance, whether the data given in Table 11.16 "Two Types of Test Strips" provide sufficient evidence to conclude that Type *B* test strips have better consistency (lower variance) than Type *A* test strips.

**\[ Solution ]**

* **Step 1.** The test of **hypotheses** is now\
  \
  &#x20;                    $$H\_0:σ\_1=σ\_2$$\
  &#x20;              vs. $$H\_a:σ\_1^2 \ne σ\_2^2   \space \space @ α=0.10$$<br>

* **Step 2.** The distribution is the ***F*****-distribution** with degrees of freedom \
  &#x20;           &#x20;

  &#x20;            $$df\_1=(16−1)=15$$ and $$df\_2=(21−1)=20$$

* **Step 3.** The **value of the test statistic** is\
  \
  &#x20;                        $$F=\frac{s^2\_1}{s^2\_2}=\frac{2.09}{1.10}=1.90$$

* **Step 4.** The test is right-tailed. The **single critical value** is \
  \
  &#x20;            $$F\_α=F\_{0.10}=1.84$$

  \
  &#x20;            Thus the **rejection region** is $$\[1.84,∞)$$, \
  &#x20;            as illustrated in Figure 11.11 "Rejection Region and Test Statistic for ".

Figure 11.11 Rejection Region and Test Statistic for [Note 11.28 "Example 7"](https://saylordotorg.github.io/text_introductory-statistics/s15-03-f-tests-for-equality-of-two-va.html#fwk-shafer-ch11_s03_s02_n03)

![](https://saylordotorg.github.io/text_introductory-statistics/section_15/0dee7b0964c04670ae90832291b81e85.jpg)

* **Step 5.** As shown in Figure 11.11 "Rejection Region and Test Statistic for ", the test statistic 1.90 lies in the rejection region, so the decision is **to reject** $$H\_0$$ . <br>
* The data provide sufficient evidence, at the 10% level of significance, to **conclude** that Type *B* test strips have better consistency (lower variance) than Type *A* test strips do.
