This preview shows pages 1–3. Sign up to view the full content.
Chapter 11
Inference for One Numerical Population
11.1
Counting
Suppose that we can observe i.i.d. random variables,
X
1
,X
2
3
,...,X
n
that are count variables; i.e. the possible values of each random variable are the integers
...,

3
,

2
,

1
,
0
,
1
,
2
,
3
,...
or some subset of the integers.
We have studied this problem: for BTs (giving the binomial); and for the Poisson. In this
section we consider the general problem.
For example, consider the population of students at UWMadison this semester with a response
equal to the total number of credits that will be completed. Personally, I would not be willing to
study this as either the binomial or the Poisson.
The general problem is as follows. The probability distribution of
X
1
is given by a collection
of equations:
P
(
X
1
=
j
)=
p
j
,
for
j
=

3
,

2
,

1
,
0
,
1
,
2
,
3
,....
The ideal situation would be when we know all of the
p
j
’s, for then we could compute the proba
bility of any event. But the ideal is not realistic in science.
The next best would be to have a
parametric family
such as the Poisson of binomial. In
these cases all we need to do is estimate one parameter (or more, sometimes a family has more
than one parameter) and then we would have estimates of all the
p
j
’s. This is a fruitful area
that we cannot pursue in this course because of time limitations. In addition to the binomial
and Poisson, parametric families include: the
geometric
,the
hypergeometric
and the
negative
binomial
distributions.
Instead, we opt for a much more modest goal. We will use our datatodrawinferencesabou
t
the mean
μ
of the probability distribution. Just as for the binomial andPoisson
,youcanvisualize
μ
as follows: Given the
p
j
’s we can draw a probability histogram; the center of gravity of the
probability histogram is the mean of the population.
127
This preview has intentionally blurred sections. Sign up to view the full version.
View Full Document We will compute the mean and standard deviation of our data, denoted by
¯
x
and
s
,asinChapter
10. We will begin with estimation.
11.1.1
Estimation of
μ
Our point estimate of
μ
,themeanofthepopu
la
t
ion
,iss
imp
ly
¯
x
,themeanofourda
ta
. Bu
t
,of
course, we want to have a conFdence interval estimate too.
It turns out that without a parametric family to guide us, an exact answer is impossible; we
have no choice but to use an approximate method. In order to obtain an approximate CI, we need
to be able to compute approximate probabilities for
¯
X
.±o
r
tuna
te
,the
reisawonde
r
fu
lre
su
l
tin
probability theory that can help us. It is called the
Central Limit Theorem (CLT).
Let’s examine this three word name. Theorem, of course, meansitisanimportantmathematical
fact. Limit means that the truth of the theorem is achieved only as
n
grows without bound. In other
words, for any Fnite value of
n
the result of the theorem is only an approximation. (Here is an
example that you might have seen in calculus. As
n
tends to inFnity, the value
1
/n
converges to 0
in the limit. ±or any Fnite
n
the limit, 0, can be viewed as an approximation to
1
/n
.) The quality
of the approximation we obtain from the CLT is an important andvex
ingissuetha
twew
i
l
ldea
l
with below. ±inally, it is called Central because it is viewedasveryimportant,i.e.central,toallof
This is the end of the preview. Sign up
to
access the rest of the document.
This note was uploaded on 12/10/2011 for the course STATS 371 taught by Professor Hanlon during the Fall '11 term at Wisconsin.
 Fall '11
 hanlon
 Counting

Click to edit the document details