0% found this document useful (0 votes)
53 views4 pages

Bayesian Inference

1) Bayesian inference is a statistical approach where unknown quantities are treated as random variables with prior distributions, which are updated using Bayes' rule after observing data to get posterior distributions. 2) Two examples are provided: estimating voter support from a sample, where the prior comes from previous elections; and estimating transmitted signals in noisy communication, where the prior is determined by the system design. 3) In both examples, the goal is to make inferences about an unobserved random variable X by observing related data Y, finding the posterior distribution of X given Y using Bayes' rule, and estimating X from the posterior.

Uploaded by

bossishere
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
53 views4 pages

Bayesian Inference

1) Bayesian inference is a statistical approach where unknown quantities are treated as random variables with prior distributions, which are updated using Bayes' rule after observing data to get posterior distributions. 2) Two examples are provided: estimating voter support from a sample, where the prior comes from previous elections; and estimating transmitted signals in noisy communication, where the prior is determined by the system design. 3) In both examples, the goal is to make inferences about an unobserved random variable X by observing related data Y, finding the posterior distribution of X given Y using Bayes' rule, and estimating X from the posterior.

Uploaded by

bossishere
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 4

Bayesian Inference https://www.probabilitycourse.com/chapter9/9_1_0_bayesian_inference.

php

9.1.0 Bayesian Inference


The following is a general setup for a statistical inference problem: There is an unknown quantity that we
would like to estimate. We get some data. From the data, we estimate the desired quantity. In the previous
chapter, we discussed the frequentist approach to this problem. In that approach, the unknown quantity θ is
assumed to be a fixed (non-random) quantity that is to be estimated by the observed data.

In this chapter, we would like to discuss a different framework for inference, namely the Bayesian approach.
In the Bayesian framework, we treat the unknown quantity, Θ , as a random variable. More specifically, we
assume that we have some initial guess about the distribution of Θ . This distribution is called the prior
distribution. After observing some data, we update the distribution of Θ (based on the observed data). This
step is usually done using Bayes' Rule. That is why this approach is called the Bayesian approach. The details
of this approach will be clearer as you go through the chapter. Here, to motivate the Bayesian approach, we
will provide two examples of statistical problems that might be solved using the Bayesian approach.

Example 9.1

Suppose that you would like to estimate the portion of voters in your town that plan to vote for Party A in an
upcoming election. To do so, you take a random sample of size n from the likely voters in the town. Since
you have a limited amount of time and resources, your sample is relatively small. Specifically, suppose that
n = 20. After doing your sampling, you find out that 6 people in your sample say they will vote for Party
A.

• Solution
◦ Let θ be the true portion of voters in your town who plan to vote for Party A. You might want to
estimate θ as

6
θ ̂  = = 0.3
20
In fact, in absence of any other data, that seems to be a reasonable estimate. However, you might
feel that n = 20 is too small. Thus, your guess is that the error in your estimation might be too
high. While thinking about this problem, you remember that the data from the previous election
is available to you. You look at that data and find out that, in the previous election, 40% of the
people in your town voted for Party A. How can you use this data to possibly improve your
estimate of θ ? You might argue as follows:

Although the portion of votes for Party A changes from one election to another, the change is not
usually very drastic. Therefore, given that in the previous election 40% of the voters voted for
Party A, you might want to model the portion of votes for Party A in the next election as a
random variable Θ with a probability density function, fΘ (θ) , that is mostly concentrated
around θ = 0.4. For example, you might want to choose the density such that

E[Θ] = 0.4

1 of 4 10/12/22, 3:53 pm
Bayesian Inference https://www.probabilitycourse.com/chapter9/9_1_0_bayesian_inference.php

Figure 9.1 shows an example of such density functions. Such a distribution shows your prior
belief about Θ in the absence of any additional data. That is, before taking your random sample
of size n = 20 , this is your guess about the distribution of Θ .

Figure 9.1 - An example of a prior distribution for Θ in Example 9.1

Therefore, you initially have the prior distribution fΘ (θ) . Then you collect some data, shown by
D. More specifically, here your data is a random sample of size n = 20 voters, 6 of whom are
voting for Party A. As we will discuss in more detail, you can then proceed to find an updated
distribution for Θ , called the posterior distribution, using Bayes' rule:

P(D|θ)fΘ (θ)
fΘ|D (θ|D) = .
P(D)
We can now use the posterior density, fΘ|D (θ|D), to further draw inferences about Θ . More
specifically, we might use it to find point or interval estimates of Θ .

Example 9.2
Consider a communication channel as shown in Figure 9.2. We can model the communication over this
channel as follows. At time n, a random variable Xn is generated and is transmitted over the channel.
However, the channel is noisy. Thus, at the receiver, a noisy version of Xn is received. More specifically, the
received signal is

Yn = Xn + Wn ,
where Wn ∼ N(0, σ 2 ) is the noise added to Xn . We assume that the receiver knows the distribution of Xn .
The goal here is to recover (estimate) the value of Xn based on the observed value of Yn .

Figure 9.2 - Noisy communication channel in Example 9.2

2 of 4 10/12/22, 3:53 pm
Bayesian Inference https://www.probabilitycourse.com/chapter9/9_1_0_bayesian_inference.php

• Solution
◦ Again, we are dealing with estimating a random variable (Xn ). In this case, the prior distribution
is fX (x). After observing Yn , the posterior distribution can be written as

fYn |Xn (y|x)fX (x)


fXn |Yn (x|y) = .
fY (y)
Here, we have assumed both X and Y are continuous random variables. The above formula is a
version of Bayes' rule. We will discuss the details of this approach shortly; however, as you'll
notice, we are using the same framework as Example 9.1. After finding the posterior distribution,
fXn |Yn (x|y), we can then use it to estimate the value of Xn .

If you think about Examples 9.1 and 9.2 carefully, you will notice that they have similar structures. Basically,
in both problems, our goal is to draw an inference about the value of an unobserved random variable (Θ or
Xn ). We observe some data (D or Yn ). We then use Bayes' rule to make inference about the unobserved
random variable. This is generally how we approach inference problems in Bayesian statistics.

It is worth noting that Examples 9.1 and 9.2 are conceptually different in the following sense: In Example
9.1, the choice of prior distribution fΘ (θ) is somewhat unclear. That is, different people might use different
prior distributions. In other words, the choice of prior distribution is subjective here. On the other hand, in
Example 9.2, the prior distribution fXn (x) might be determined as a part of the communication system
design. In other words, for this example, the prior distribution might be known without any ambiguity.
Nevertheless, once the prior distribution is determined, then one uses similar methods to attack both
problems. For this reason, we study both problems under the umbrella of Bayesian statistics.

Bayesian Statistical Inference

The goal is to draw inferences about an unknown variable X by observing a related random variable Y . The
unknown variable is modeled as a random variable X , with prior distribution

fX (x), if X is continuous,


PX (x), if X is discrete.
After observing the value of the random variable Y , we find the posterior distribution of X . This is the
conditional PDF (or PMF) of X given Y = y,

fX|Y (x|y)  or  PX|Y (x|y).


The posterior distribution is usually found using Bayes' formula. Using the posterior distribution, we can then
find point or interval estimates of X .
Note that in the above setting, X or Y (or possibly both) could be random vectors. For example,
X = (X1 , X2 , ⋯ , Xn ) might consist of several random variables. However, the general idea of Bayesian
statistics stays the same. We will specifically talk about estimating random vectors in Section 9.1.7.
← previous
next →

3 of 4 10/12/22, 3:53 pm
Bayesian Inference https://www.probabilitycourse.com/chapter9/9_1_0_bayesian_inference.php

The print version of the book is available through Amazon here.

4 of 4 10/12/22, 3:53 pm

You might also like