J.Opet- Senior Composition

38
The Central Limit Theorem and the Theory of Hypothesis Testing, with an Application to Rental Income from Two Populations A Senior Comprehensive Project by Johnathon M. Opet Allegheny College Meadville, PA December 8, 2015 Submitted to the Department of Mathematics in partial fulfillment of the requirements for the degree of Bachelor of Science. Project Advisor: Dr. Anthony Lo Bello Second Reader: Dr. Rachel Weir I hereby recognize and pledge to fulfill my responsibilities, as defined in the Honor Code, and to maintain the integrity of both myself and the College community as a whole. Pledge: —————————————————————– Johnathon M. Opet

Transcript of J.Opet- Senior Composition

The Central Limit Theorem andthe Theory of Hypothesis Testing,

with an Application to RentalIncome from Two Populations

A Senior Comprehensive Projectby

Johnathon M. OpetAllegheny College

Meadville, PA

December 8, 2015

Submitted to the Department of Mathematics in partial fulfillment ofthe requirements for the degree of Bachelor of Science.

Project Advisor: Dr. Anthony Lo BelloSecond Reader: Dr. Rachel Weir

I hereby recognize and pledge to fulfill my responsibilities, as definedin the Honor Code, and to maintain the integrity of both myself and

the College community as a whole.

Pledge:

—————————————————————–Johnathon M. Opet

Acknowledgements

I want to thank my parents, Dee and John Opet, for all the hard work they

have done for me and making this opportunity a reality. You two have taught

me that things are never handed but are deserved through hard work. You

have shown me how to become a man and I could never thank you both enough

for what you have taught me and what you have given me. I love you both,

thank you for giving me the world.

Abstract

This senior project presents the proof of the Central Limit Theorem

and a discussion of the theory of Hypothesis Testing, and concludes with

some applications to rental incomes of college and residential houses in

Northeast Ohio and Northwest and Southwest Pennsylvania.

Contents

1 Preliminary Definitions and Theorems 1

2 The Central Limit Theorem 9

3 The Theory of Hypothesis Testing 20

4 Application: Rental Income from Two

Populations 25

1 Preliminary Definitions and Theorems

Definition 1.1. A Sample Space, X, is a nonempty set of points, called

outcomes.

Definition 1.2. A Sigma Field of Events, A, is a non-empty collection of

subsets of X satisfying two conditions:

1. If A ∈ A then A′ ∈ A

2. If A1, A2, . . . , An, · · · ∈ A then∞⋃i=1

Ai ∈ A.

The elements of A are called events.

Theorem 1. If A is a sigma field of events of a sample space, X, then X ∈ A.

Proof. Since A is nonempty, there exists an event A ∈ A. Then we know that

A′ ∈ A. Therefore A ∪ A′ ∈ A, but A ∪ A′ = X, so X ∈ A.

Corollary 2. If A is a sigma field of events of a sample space, X, then ∅ ∈ A

Proof. Since X ∈ A, X′ ∈ A. But X′ = ∅, so ∅ ∈ A.

Definition 1.3. Probability, P, is a function with domain A and range in

the real numbers that satisfies the following conditions:

1. P (A) ≥ 0 for all events A.

2. P (X) = 1 where X is the sample space.

3. If A1, A2, ... is a mutually disjoint sequence of events, then

P

(∞⋃i=1

Ai

)=∞∑i=1

P (Ai)

.

1

Definition 1.4. A random variable X is a function from the sample space X

to the set of real numbers.

Definition 1.5. A Discrete Random Variable X is a random variable

whose range is finite or countably infinite. A Continuous Random Variable

X is a random variable whose range is an interval of length greater than 0.

Definition 1.6 ([3]). If X is a discrete random variable, the function given

by f(x) = P (X = x) for each x with the range of X is called the probability

function of X.

Definition 1.7 ([3]). A function, f(x), from the set of real numbers to the set

of real numbers is called a probability density function of the continuous

random variable X if and only if

P (a ≤ X ≤ b) =

∫ b

a

f(x)dx

for any real constants a and b with a ≤ b.

Definition 1.8 ([3]). The rth Moment about the origin of a random

variable X is

µ′r = E(Xr) =∑x

xr · f(x), r = 0, 1, 2, . . . ,

when X is discrete, and

µ′r = E(Xr) =

∫ ∞−∞

xr · f(x)dx, r = 0, 1, 2, . . . ,

2

when X is continuous.

Definition 1.9 ([3]). µ′1 is the mean or expected value of the random

variable X and is denoted by µ or E(X).

Definition 1.10. The rth moment about the mean of the random variable

X is

µr = E[(X− µ)r] =∑x

(x− µ)r · f(x), r = 0, 1, 2, . . . ,

when X is discrete, and

µr = E[(X− µ)r] =

∫ ∞−∞

(x− µ)r · f(x)dx, r = 0, 1, 2, . . . ,

when X is continuous.

Definition 1.11. µ2 is the variance of the random variable X and is denoted

by σ2 or var(X). The positive square root, σ, of the variance is the standard

deviation.

Theorem 3.

σ2 = µ′2 − µ2

for any random variable X with mean µ and variance σ2.

3

Proof.

σ2 =E[(X− µ)2

]=E(X2 − 2µx+ µ2)

=E(X2)− 2µE(X) + E(µ2)

=E(X2)− 2µ · µ+ µ2

=µ′2 − 2µ2 + µ2

=µ′2 − µ2.

Definition 1.12. If X1,X2, . . . ,Xn are discrete random variables, the func-

tion given by f(x1, x2, . . . , xn) = P (X1 = x1,X2 = x2, . . . ,Xn = xn) for each

point (x1, x2, . . . , xn), where xi is in the range of Xi, is the joint distribution

of X1,X2, . . . ,Xn.

Definition 1.13. If X1,X2, . . . ,Xn are continuous random variables, the

function F (x1, x2, . . . , xn), defined by F (x1, x2, . . . , xn) = P (X1 < x1, . . . ,Xn <

xn), is the joint cumulative distribution function of X1,X2, . . . ,Xn and

the function f(x1, . . . , xn) = ∂nF∂xn···∂x2∂x1 is the joint probability density function

of X1,X2, . . . ,Xn.

Definition 1.14 ([3]). If f(x1, x2, . . . , xn) is the value of the joint probability

distribution of the n random variables X1,X2, . . . ,Xn at (x1, x2, . . . , xn), and

fi(xi) is the value of the marginal distribution of Xi at xi for i = 1, 2, . . . , n,

4

then the n random variables are independent if and only if

f(x1, x2, . . . , xn) = f1(x1) · f2(x2) · · · fn(xn)

for all (x1, x2, . . . , xn) where xi is within the range of Xi.

Theorem 4 (Linear Property of Expectations).

E(aX + bY) = aE(X) + bE(Y)

Proof. We give the proof for the continuous case. Suppose the random vari-

ables X and Y have probability density functions g(x) and h(y) respectively,

and let f(x, y) be the joint probability density function of X and Y. Then

E(aX + bY) =

∫ ∞−∞

∫ ∞−∞

(ax+ by)f(x, y)dydx

=a

∫ ∞−∞

∫ ∞−∞

xf(x, y)dydx+ b

∫ ∞−∞

∫ ∞−∞

yf(x, y)dxdy

=a

∫ ∞−∞

x

(∫ ∞−∞

f(x, y)dy

)dx+ b

∫ ∞−∞

y

(∫ ∞−∞

f(x, y)dx

)dy

=a

∫ ∞−∞

xg(x)dx+ b

∫ ∞−∞

yh(y)dy

=aE(X) + bE(Y).

Definition 1.15. If X and Y are discrete random variables with joint prob-

5

ability function f(x, y), then the covariance of X and Y is

cov(X,Y) =∑y

∑x

(x− E(X))(y − E(Y))f(x, y).

If X and Y are continuous random variables with joint probability density

function f(x, y), then

cov(X,Y) =

∫ ∞−∞

∫ ∞−∞

(x− E(X))(y − E(Y))f(x, y)dydx.

Theorem 5.

V ar(aX + bY) = a2σ2X + 2ab cov(X,Y) + b2σ2

Y

Proof. We give the proof for the continuous case. Suppose the random vari-

ables X and Y have probability density functions g(x) and h(y) respectively,

and let f(x, y) be the joint probability density function of X and Y. Then

V ar(aX + bY) =

∫ ∞−∞

∫ ∞−∞

[(ax+ by)− (aE(X) + bE(Y))]2f(x, y)dydx

=

∫ ∞−∞

∫ ∞−∞

[ax− aE(X) + by − bE(Y)]2f(x, y)dydx

=

∫ ∞−∞

∫ ∞−∞

[a(x− E(X)) + b(y − E(Y))]2f(x, y)dydx

=a2∫ ∞−∞

∫ ∞−∞

(x− E(X))2f(x, y)dydx

+ 2ab

∫ ∞−∞

∫ ∞−∞

(x− E(X))(y − E(Y))f(x, y)dydx

+ b2∫ ∞−∞

∫ ∞−∞

(y − E(Y))2f(x, y)dydx

6

=a2∫ ∞−∞

(x− E(X))2[ ∫ ∞−∞

f(x, y)dy

]dx+ 2abcov(X,Y)

+ b2∫ ∞−∞

(y − E(Y))2[ ∫ ∞−∞

f(x, y)dx

]dy

=a2∫ ∞−∞

(x− E(X))2g(x)dx+ 2abcov(X,Y)

+ b2∫ ∞−∞

(y − E(Y))2h(y)dy

=a2var(X) + 2ab cov(X,Y) + b2var(Y)

Theorem 6. If X and Y are independent random variables, then cov(X,Y) =

0.

Proof. For the continuous case, if f(x, y), g(x), and h(y) are the same functions

as in the previous theorem, then

cov(X,Y) =

∫ ∞−∞

∫ ∞−∞

(x− E(X))(y − E(Y))f(x, y)dydx

=

∫ ∞−∞

∫ ∞−∞

(x− E(X))(y − E(Y))g(x)h(y)dydx

=

∫ ∞−∞

(x− E(X))g(x)dx

∫ ∞−∞

(y − E(Y))h(y)dy

=0 · 0 = 0.

Theorem 7. If X1,X2, . . .Xn is a sequence of independent random variables

with common mean, µ, and common variance, σ2, and if X = X1+X2+···+Xn

n,

7

then E(X) = µ and var(X) = σ2

n.

Proof. This theorem follows from Theorems 4, 5, and 6.

Definition 1.16. [1] X has Normal Distribution with parameters µ and σ

if it has probability density function given by

n(x;µ, σ) =1

σ√

2πe

12(x−µσ

)2

where −∞ < x <∞ and σ > 0.

Theorem 8.

ln(1 + x) = x− x2

2+x3

3− x4

4+− · · ·

for |x| < 1.

Proof. If |x| < 1, then 11+x

= 1− x+ x2 − x3 +− · · · .

Therefore, for |x| < 1

ln(1 + x) =

∫ x

0

1

1 + tdt

=

∫ x

0

(1− t+ t2 − t3 + · · · )dt

=x− x2

2+x3

3− x4

4+− · · · .

8

2 The Central Limit Theorem

Before we introduce the Central Limit Theorem, we must first understand

the background mathematics behind the proof of the theorem. First, we will

introduce the moment generating function.

Definition 2.1 ([3]). The moment generating function of a random vari-

able X, where it exists, is given by

MX(t) = E(etX) =∑x

etx · f(x)

when X is discrete and f is its probability function and

MX(t) = E(etX) =∫∞−∞ e

tx · f(x)dx

when X is continuous and f is its probability density function.

In addition, we will also need to introduce a property of moment generating

functions.

Theorem 9.

MX(t) = 1 + µt+ µ′2t2

2!+ µ′3

t3

3!+ · · · (1)

Proof. Let X be a continuous random variable with probability density func-

tion f(x) and let Y = eXt. Then

MX(t) = E(Y) =

∫ ∞−∞

yf(x)dx

=

∫ ∞−∞

extf(x)dx

9

Since we know the expansion of the Maclaurin Series for eu is eu =∑∞

n=0un

n!,

substituting u = xt we get

E(Y) =

∫ ∞−∞

∞∑n=0

(xt)n

n!f(x)dx

=∞∑n=0

tn

n!

∫ ∞−∞

xnf(x)dx.

Since we know∫∞−∞ x

nf(x)dx = µ′n we get

E(Y) =∞∑n=0

µ′ntn

n!.

Theorem 10. If a and b are constants; then

MaX+b(t) = ebtMX(at). (2)

Proof. We will give the proof for the case when X is a continuous random

variable. Let a and b be constants, then

10

MaX+b(t) =

∫ ∞−∞

e(ax+b)tf(x)dx

=

∫ ∞−∞

eaxtebtf(x)dx

= ebt∫ ∞−∞

exatf(x)dx

= ebtMX(at).

For example, if we wanted to find the moment generating function of the

random variable Y = 2X + 8, we could use this theorem to get

M2X+8 = e8tMX(2t).

Another tool we will need in order to prove the Central Limit Theorem is:

Theorem 11. If X1,X2, . . . , and Xn are independent random variables and

Y = X1 + X2 + · · ·+ Xn, then

MY(t) =n∏i=1

MXi(t) (3)

where MXi(t) is the value of the moment-generating function of Xi at t.

Proof. For the discrete case, let X1,X2, . . . , and Xn be independent random

variables and Y = X1 + X2 + · · ·+ Xn. Then we know

11

f(x1, x2, . . . , xn) = f1(x1) · f2(x2) · ... · fn(xn).

Then we can write

MY(t) = E(eYt)

= E[e(X1+X2+···+Xn)t]

=∑x1

· · ·∑xn

e(x1+x2+···+xn)tf(x1, x2, . . . , xn)

=∑x1

ex1tf1(x1) ·∑x2

ex2tf2(x2) · · ·∑xn

exntfn(xn)

=n∏i=1

MXi(t)

This theorem states that the moment-generating function of the sum of

n independent random variables is the product of their moment-generating

functions. As in the case of Theorem 10, this theorem allows us to manipulate

the moment generating function into a more simple form to work with.

Theorem 12. If two random variables have the same moment generating func-

tion, then they have the same probability distribution.

Proof. Omitted

12

Theorem 13. If X1,X2, . . . , are random variables such that

limn→∞

MXn(t) = MX(t)

then

limn→∞

P (Xn < a) = P (X < a)

for all real numbers a.

Proof. Omitted

We now proceed to discuss how we estimate an unknown population vari-

ance σ2 when we need such an estimate to do a hypothesis test.

Definition 2.2 ([3]). Let X1,X2, . . . ,Xn be a random sample from a distri-

bution having unknown finite variance σ2. Then

S2 =n∑i=1

(Xi −X)2

n− 1(4)

is called the sample variance of X1,X2, . . . ,Xn.

Theorem 14 ([3]). E(S2) = σ2

13

Proof. Using Equation (4) to calculate the expected value, we get

E(S2) =E

[ n∑i=1

(Xi −X)2

n− 1

]=

1

n− 1E

[ n∑i=1

(Xi −X)2]

=1

n− 1E

[ n∑i=1

(Xi2 − 2XiX + X

2)

]=

1

n− 1E

[ n∑i=1

Xi2 − 2X

n∑i=1

Xi + nX2]

=1

n− 1E

[ n∑i=1

Xi2 − 2nX

2+ nX

2]

=1

n− 1E

[ n∑i=1

Xi2 − nX

2]

=1

n− 1

[ n∑i=1

E(Xi2)− nE(X

2)

]

Since E(X2i ) = µ′2, and µ′2 = σ2 + µ2, and since E(X

2) = σ2

n+ µ2, we get

E

[ n∑i=1

(Xi −X)2

n− 1

]=

1

n− 1

[ n∑i=1

E(Xi2)− nE(X

2)

]=

1

n− 1[n(σ2 + µ2)− n(

σ2

n+ µ2)]

=1

n− 1[nσ2 + nµ2 − σ2 − nµ2)]

=1

n− 1[(n− 1)σ2]

=σ2.

14

Because of this theorem, we call S2 an unbiased estimator of σ2. Its

value s2 is called the estimate of σ2. We use the value of this unbiased

estimator, S2, to estimate σ2 from data when σ2 is unknown.

Theorem 15. The moment generating function of a random variable X with

normal distribution with parameters µ and σ is eµt+t2σ2

2 .

Proof.

MX(t) =

∫ ∞−∞

1√2πσ

e−12((x−µ)/σ)2extdx

=1√2πσ

∫ ∞−∞

ext−12((x−µ)/σ)2dx.

Looking at the exponent, xt− 12(x−µ

σ)2 we can see that,

xt− 1

2

(x− µσ

)2

=− 1

2σ2[−2xtσ2 + x2 − 2xµ+ µ2]

=− 1

2σ2[x2 − (2µ+ 2tσ2)x+ µ2]

=− 1

2σ2[x2 − (2µ+ 2tσ2)x+ (µ+ tσ2)2 − (µ+ tσ2)2 + µ2]

=− 1

2σ2[x− (µ+ tσ2)]2 +

1

2σ2[(µ+ tσ2)2 − µ2]

=− 1

2

[x− (µ+ tσ2)

σ

]2+

1

2σ2[µ2 + 2µtσ2 + t2σ4 − µ2]

=− 1

2

[x− (µ+ tσ2)

σ

]2+ µt+

t2σ2

2.

15

So,

MX(t) =

∫ ∞−∞

1√2πσ

ext−12((x−µ)/σ)2dx

=eµt+t2σ2

2

∫ ∞−∞

1√2πσ

e−12

[x−(µ+tσ2)

σ

]2dx

=eµt+t2σ2

2 .

If µ = 0 and σ = 1, then X has standard normal distribution and the

moment generating function is

MX(t) =et2

2 .

Theorem 16. If X and Y are independent random variables with normal

distribution with parameters µX, σ2X, and µY, σ2

Y respectively, then aX + bY

has a normal distribution with parameters µ = aµX + bµY and σ2 = a2σ2X +

b2σ2Y.

Proof. This result follows from Theorems 10, 11, 12, and 15.

The Central Limit Theorem is one of the fundamental theorems of probabil-

ity. What the Central Limit Theorem states is that the standardized average

value of a large sample of independent random variables will have approxi-

mately standard normal distribution regardless of any other aspects like the

underlying distribution [3].

16

Theorem 17 ([2], The Central Limit Theorem). Let X1,X2, . . . ,Xn, . . .

be independent and identically distributed random variables with mean µ and

standard deviation σ. Let X be the sample mean with sample size n. Then, as

n approaches infinity, the distribution of

Z =X− µ√σ2/n

, (5)

approaches that of the standard normal distribution.

Proof. If we refer back to Equation (10), we get

MZ(t) = M X−µσ/√n

(t) =M√nσ

X−√nµσ

(t)

=e−√nµt/σ ·MX

(√nt

σ

)=e−

√nµt/σ ·MnX

(t

σ√n

)

Since we know nX = X1 +X2 + · · ·+Xn, and since the Xi’s have the same

distribution, then it follows from Equation (3) that

MZ(t) = e−√nµt/σ ·

[MX

(t

σ√n

)]n

where X is a random variable with the same distribution as the Xi’s.

If we take the natural log of each side we obtain,

17

lnMZ(t) = −√nµt

σ+ n · ln

[MX

(t

σ√n

)]

by the properties of logarithms. If we expand MX

(t

σ√n

)as a power series in

t, we obtain

lnMZ(t) = −√nµt

σ+ n · ln

[1 + µ′1

t

σ√n

+ µ′2t2

2σ2n+ µ′3

t3

6σ3n√n

+ · · ·]

where µ′1, µ′2, µ

′3, . . . , are the moments about the origin of the population

distribution, that is, those of the original random variables Xi.

We can use the infinite series ln(1 + x) = x− 12x2 + 1

3x3−+ · · · for |x| < 1

to get

lnMZ(t) =−√nµt

σ+ n

{[µ′1

t

σ√n

+ µ′2t2

2σ2n+ µ′3

t3

6σ3n√n

+ · · ·]

− 1

2

[µ′1

t

σ√n

+ µ′2t2

2σ2n+ µ′3

t3

6σ3n√n

+ · · ·]2

+1

3

[µ′1

t

σ√n

+ µ′2t2

2σ2n+ µ′3

t3

6σ3n√n

+ · · ·]3 −+ · · ·

}.

Then, collecting the powers of t, we obtain

18

lnMZ(t) =

(−√nµ

σ+

√nµ′1σ

)t+

(µ′22σ2− µ′1

2

2σ2

)t2

+

(µ′3

6σ3√n− µ′1 · µ′2

2σ3√n

+µ′1

3

3σ3√n

)t3 + · · ·

and since µ′1 = µ and σ2 = µ′2 − µ2 so µ′2 = σ2 + µ2, we can reduce this to

lnMz(t) =1

2t2 +

(µ′36− µ′1µ

′2

2+µ′1

3

6

)t3

σ3√n

+ · · ·

Finally, observing that the coefficient of t3 is a constant times 1√n

and in

general, for r ≥ 2, the coefficient of tr is a constant times 1√nr−2

, we get

limn→∞

lnMZ(t) =1

2t2.

Thus

limn→∞

MZ(t) = et2/2,

which is the moment generating function of the random variable with standard

normal distribution.

19

3 The Theory of Hypothesis Testing

Definition 3.1. [3] A statistical hypothesis is an assertion or conjecture

about the distribution of one or more random variables. If a statistical hy-

pothesis completely specifies the distribution, it is referred to as a simple

hypothesis; if not, it is referred to as a composite hypothesis.

In order to be able to construct the correct criteria for testing statistic

hypotheses, we must first form our Null Hypothesis, H0, and our Alterna-

tive hypothesis, H1. The null hypothesis is the claim that will be tested and

assumed true until proven unlikely by the data. In contrast, the alternative

hypothesis is always a claim that denies the null hypothesis. After formulating

the two hypotheses for our test, there are two types of errors that can possibly

occur. The table below represents the relationship of our claim of H0 and the

truth or falsity of the hypothesis formed.

Do Not Reject H0 Reject H0

H0 is True Correct Type I ErrorH1 is True Type II Error Correct

Table 1: [2] Type I & II Error Table

When testing our hypothesis, we must be careful to avoid errors, which are

labeled in red. A Type I error is an error when we reject H0 when it is in

fact true, and a Type II error is an error when we do not reject H0 when it

is false.

20

Definition 3.2. α, the level of significance, is the probability of commit-

ting a Type I error, and β is the probability of committing a Type II error.

Since a Type I error is considered more serious than a Type II error, we

want to reduce that probability by selecting a small value for α. If rejecting the

null hypothesis would cause the closing down of a business, we would select a

very small value for α like .01 or .001 to reduce the risk of committing a Type

I error. If there is no terrible consequence associated with rejecting the null

hypothesis, then we would choose a slightly larger α like .1 or .05. Once we

have picked our α, we are able to generate our critical values for a two-tailed

test.

Definition 3.3. The decision whether to accept or reject the null hypothesis

is made based on the test statistic value, z, of a random variable Z. The

numbers −zα2

and zα2

are the critical values of Z. If −zα2< z < −zα

2, then we

accept the null hypothesis, whereas if zα/2 < z or z < −zα2, then we reject the

null hypothesis.

In our example, we will be concerned with the situation where our decision

regarding H0 depends on the value of z of a random variable Z with standard

normal distribution

z =x− µσ/√n

(6)

where x is the value of the sample mean of size n from a population with mean

µ and variance σ2. The test statistic, z, can be located in the non-rejection

region or in the rejection region. Graphically this is represented below:

21

Looking at the graph, if the test statistic is located in (−∞,−zα/2) or

(zα/2,∞), then it is in the rejection region, which means that we must not

accept H0. If the test statistic is located in the interval (−zα/2, zα/2), then it is

in the non-rejection region, which means that we must accept H0. If the test

statistic is either equal to, or relatively close to the critical values, then it is

up to the experimenter to decide whether to defer judgment.

Example 3.1. Suppose that it is known from experience that the standard

deviation of the weight of 8-ounce packages of cookies made by a certain bakery

is 0.16 ounces. To check whether its production is under control on a given

day, that is, to check whether the true average weight of the packages is 8

ounces, employees select a random sample of 25 packages and find that their

mean weight is x = 8.091 ounces. Is the bakery telling the truth when they

state that their package of cookies are 8-ounces? [3]

Solution: The null hypothesis is that the package of cookies does weigh

8-ounces, in other words, H0 : µ = 8. In contrast, the alternative hypothesis is

that the package of cookies does not weight 8-ounces, in other words, H1 : µ 6=

8. Since the application is associated with a lot of risk (if the null hypothesis

22

is rejected the bakery must stop production), we will choose a low level of

significance namely, α = 0.01.

Now that we have found our value for α, we must find the critical values.

When α = .005 and n = 25, the critical values are z.005 = 2.575 and −z.005 =

−2.575.

After finding the critical values, we must calculate the test statistic with

the given data. Substituting n = 25, σ = 0.16, x = 8.091, and µ = 8 into

Equation (6), we obtain

z =8.091− 8

0.16/√

25= 2.844.

This can be represented graphically below:

Looking at the graph, we know that the rejection region consists of the union

of the two intervals (−∞,−2.575) and (2.575,∞). Since our critical value

is greater than 2.575, the null hypothesis must be rejected. Therefore it is

highly unlikely that the 8-ounce packages of cookies actually weigh 8-ounces.

Therefore, the bakery must shut down production of their packages of cookies

and fix the problem immediately.

23

Example 3.2. The owner of Tony’s Tires claims that his manufactured tires

last on average 22,000 miles. We wanted to test whether or not Tony was

telling the truth. Using the data collected, what can we say about his claim?

Solution: We must first produce our null hypothesis and alternative hy-

pothesis. Based on the information given, the null hypothesis H0 : µ = 22, 000

miles and the alternative hypothesis H1 : µ 6= 22, 000 miles. We choose

α = .01. After collecting and analyzing the data, we were able to obtain

the information that n = 100, x = 21, 800 miles, and s = 1, 295. Since we

know n ≥ 30, we can apply the Central Limit Theorem and the z-statistic test

to obtain

z =x− µs/√n

=21, 800− 22, 000

1295/√

100

≈ −1.54.

We know that the test statistic is -1.54. Since α = .01, we want critical

value for z.005 and −z.005. Looking Table IV in [3], we get a value of z.005 =

2.575 and −z.005 = −2.575 [3]. So we know that our rejection region is the

union of the intervals (−∞,−2.575) and (2.575,∞), and our non-rejection

region is (−2.575, 2.575). Since −2.575 < −1.54 < 2.575, we must accept our

null hypothesis, which means that the data in this case do not contradict the

owner’s claim.

24

4 Application: Rental Income from Two

Populations

Example 4.1. I was interested in finding whether or not college students living

in privately owned houses and non-college students living in privately owned

houses had the same amount of rental income. In order to obtain the data, I

sent out emails and called multiple landlords around Northeast Ohio and both

Northwest and Southwest Pennsylvania to collect data for each population. I

then formed the statistical question:

“Is the mean rental income obtained from college students living

in privately owned houses the same as the mean rental income ob-

tained from non-college students living in privately owned houses?”

I gathered up a random total of n = 33 family rentals and m = 34 college

rentals for a total of 67 total data samples, which are shown below:

Area Tenant Type Rental Income Duration Yearly Income

NE OH Non-College $1,200.00 Monthly $14,400.00

SW PA Non-College $425.00 Monthly $5,100.00

SW PA Non-College $425.00 Monthly $5,100.00

SW PA Non-College $300.00 Monthly $3,600.00

SW PA Non-College $300.00 Monthly $3,600.00

SW PA Non-College $375.00 Monthly $4,500.00

SW PA Non-College $475.00 Monthly $5,700.00

SW PA Non-College $620.00 Monthly $7,440.00

25

SW PA Non-College $470.00 Monthly $5,640.00

NE OH Non-College $2,000.62 Monthly $24,007.44

NE OH College $1,688.76 Monthly $20,265.12

NE OH College $1,688.75 Monthly $20,265.00

NE OH College $1,774.43 Monthly $21,293.16

NE OH College $1,774.43 Monthly $21,293.16

NE OH College $2,177.58 Monthly $26,130.96

NE OH College $1,700.41 Monthly $20,404.92

NW PA College $935.00 Monthly $11,220.00

NE OH College $1,658.32 Monthly $19,899.84

NE OH College $1,637.35 Monthly $19,648.20

NE OH College $1,950.88 Monthly $23,410.56

NE OH College $1,664.83 Monthly $19,977.96

NE OH College $2,269.19 Monthly $27,230.28

NE OH College $2,076.89 Monthly $24,922.68

NE OH College $2,009.14 Monthly $24,109.68

NE OH College $2,500.00 Monthly $30,000.00

NE OH College $2,000.00 Monthly $24,000.00

NW PA College $980.00 Monthly $11,760.00

NW PA College $2,100.00 Monthly $25,200.00

NW PA College $1,650.00 Monthly $19,800.00

NW PA College $1,200.00 Monthly $14,400.00

NE OH College $1,299.00 Monthly $15,588.00

NE OH College $2,000.00 Monthly $24,000.00

26

NE OH College $1,450.00 Monthly $17,400.00

NE OH College $1,499.00 Monthly $17,988.00

NE OH College $1,750.00 Monthly $21,000.00

NE OH College $1,250.00 Monthly $15,000.00

NW PA College $1,980.00 Monthly $23,760.00

NW PA College $980.00 Monthly $11,760.00

NW PA College $1,040.00 Monthly $12,480.00

NW PA College $1,130.00 Monthly $13,560.00

SW PA Non-College $2,100.00 Monthly $25,200.00

SW PA Non-College $625.00 Monthly $7,500.00

SW PA Non-College $925.00 Monthly $11,100.00

SW PA Non-College $900.00 Monthly $10,800.00

SW PA Non-College $545.00 Monthly $6,540.00

SW PA Non-College $725.00 Monthly $8,700.00

SW PA Non-College $3,500.00 Monthly $42,000.00

SW PA Non-College $2,200.00 Monthly $26,400.00

NE OH Non-College $1,100.00 Monthly $13,200.00

NE OH Non-College $1,200.00 Monthly $14,400.00

NE OH Non-College $1,150.00 Monthly $13,800.00

NE OH Non-College $1,300.00 Monthly $15,600.00

SW PA Non-College $1,130.00 Monthly $13,560.00

SW PA Non-College $985.00 Monthly $11,820.00

SW PA Non-College $1,100.00 Monthly $13,200.00

SW PA Non-College $1,350.00 Monthly $16,200.00

27

SW PA Non-College $1,230.00 Monthly $14,760.00

SW PA Non-College $1,005.00 Monthly $12,060.00

SW PA Non-College $1,230.00 Monthly $14,760.00

SW PA Non-College $1,130.00 Monthly $13,560.00

NE OH College $1,756.00 Monthly $21,072.00

NE OH College $10,850.00 Yearly $10,850.00

NE OH College $11,000.00 Yearly $11,000.00

NE OH College $12,000.00 Yearly $12,000.00

NE OH Non-College $18,360.00 Yearly $18,360.00

NW PA Non-College $13,560.00 Yearly $13,560.00

NW PA Non-College $8,100.00 Yearly $8,100.00

Table 2: Data

After multiple calculations, the summary of my results is below.

Family Rentals College Rentals

x = $12,856.59 y = $19,196.75s2x = 62,023,794.83 s2y = 28,063,536.74σx = 7,785.52 σy = 5,297.50

Table 3: Population Information

I decided to take α = .10 since there are no serious consequences if the null

hypothesis is rejected. Is there enough evidence when α = .10 to conclude that

28

the rental income from non-college students is the same as the rental income

from college students when renting from privately owned houses?

Solution: We must first formulate our null hypothesis in order to test for

the equality of means of the two populations. Thus, we want to test the null

hypothesis, which is

Ho : µX = µY

or that the means of both populations are equal, against the alternative hy-

pothesis:

H1 : µX 6= µY

in other words, that the means are not equal.

Since n and m are both greater than 30, we know by Theorem 16 that

Z = X−Y−(µx−µy)√(σ2x/n)+(σ2

y/m)has approximately standard normal distribution. In our

example, when we plug the desired values, assuming that µx = µy, we obtain

a test-statistic value,

z =(12, 856.59− 19, 196.75)− (0)√

(62, 023, 794.83/33) + (28, 063, 536.74/34)= −3.855.

29

When α = .10, we are able to get the critical z, zα/2 = z.05 = 1.645

from Table III [3]. Thus we know that our rejection regions are the intervals

(−∞,−1.645) and (1.645,∞). So looking back at the z = −3.8550, what we

find is that the test-statistic lies in the rejection region; therefore we must

reject H0. This finding implies that the means of the two populations are not

the same.

30

Example 4.2. Looking back at Example 4.1, suppose I wanted to compare

the overall rental income between renting in Northeast Ohio and Western

Pennsylvania. I then formed a different statistical question:

“Is the overall rental income obtained from Northeast Ohio the

same as the overall rental income obtained from Western Pennsyl-

vania?”

After the reorganization of the data, I gathered up a random total of n = 32

Northeastern Ohio rentals and m = 35 Western Pennsylvania rentals for a total

of again the 67 total data samples in Example 4.1. After multiple calculations,

the summary of my results is below.

Northeast Ohio Western Pennsylvania

x = $19,453.66 y = $12,984.00s2x = 66,995,576.56 s2y = 105,368,783.63σx = 8,185.08 σy = 10,264.93

Table 4: Application 2: Population Information

I decided again to take α = .10 for the same reasons as in the previous

example. Is there enough evidence when α = .10 to conclude that the rental

income from Northeast Ohio is the same as from Western Pennsylvania?

Solution: Similarly as in the first example, we must first formulate our

null hypothesis in order to test for the equality of means of the two populations.

31

Thus, we want to test the null hypothesis, which is

Ho : µX = µY

or that the means of both populations are equal, against the alternative hy-

pothesis:

H1 : µX 6= µY

in other words, that the means are not equal.

Since n and m are both greater than 30, we know that Z = X−Y−(µx−µy)√(σ2x/n)+(σ2

y/m)

has approximately standard normal distribution. In our example, when we

plug the desired values, assuming that µx = µy, we obtain a test-statistic

value,

z =(19,453.66− 12,984.00)− (0)√

(66,995,576.56/32) + (105,368,783.63/35)= 2.864.

When α = .10, we are able to get the critical z, zα/2 = z.05 = 1.645

32

from Table III [3]. Thus we know that our rejection regions are the intervals

(−∞,−1.645) and (1.645,∞). So looking back at the z = 2.864, what we

find is that the test-statistic lies in the rejection region; therefore we must

reject H0. This finding implies that the means between the two populations

are again not the same.

33

References

[1] Broverman, Samuel. ACTEX P/1 Study Manual. 2012 edition, Winsted:

ACTEX Publications, Inc., 2012.

[2] Lee, Cheng F., and John C. Lee. Statistics for Business and Financial

Economics. 3rd edition, New York: Springer, 2013.

[3] Miller, Irwin, and Marylees Miller.John E. Freund’s Mathematical Statis-

tics. 6th edition, Upper Saddle River: Prentice Hall, 1999.

34