MGT-Three
MGT-Three
2
❑ But economic theory is seldom so simple for,
besides income, a number of other variables are
also likely to affect consumption expenditure.
❑ An obvious example is wealth of the consumer.
❑ Similarly, the number of cars sold might plausibly
depend on the
1. price of cars
2. price of public transport
3. price of petrol
4. extent of the public’s concern about global
warming 3
❑ Therefore, we need to extend our simple two-
variable regression model to cover models
involving more than two variables.
Yi = 0 + 1 X 1i + 2 X 2i + • • • + K X Ki + i
Residual
Dependent (Response) Independent/Explanatory
5
variable (for sample) variables (for sample)
What changes as we move from simple to
multiple regression?
1. Potentially more explanatory power with
more variables;
2. The ability to control for other variables;
(and the interaction of various explanatory
variables: correlations and multicollinearity);
3. Harder to visualize drawing a line through
three or more (n)-dimensional space.
4. The R2 is no longer simply the square of the
correlation coefficient b/n Y and X.
6
Slope ( j ):
Ceteris paribus, Y changes by units for every
1 unit change in X j, on average.
Y-Intercept ( 0 ): j
8
❑ Further Explanations of the Assumptions:
1) The error term has zero mean E(ui) = 0
❑ This means that for each value of x, the random
variable(u) may assume various values, some
greater than zero and some smaller than zero.
❑ But if we considered all the positive and
negative values of u, for any given value of X,
they would have on average value equal to zero.
❑ The positive and negative values of u cancel
each other.
❑ Mathematically: E (U i ) = 0
Or E(ɛi|Xji) = 0. (for all i = 1, 2, …, n; j = 1, …, K)
9
2).The error term has a constant variance
(Homoscedasticity)
❑ Var (𝑢𝑖 ) = 𝐸 𝑢𝑖 2 = 𝛿 2 ) 𝑓𝑜𝑟 𝑎𝑙𝑙 𝑖 or var(ɛi|Xji) = σ2.
❑ For all values of X, the u’s will show the same
dispersion around their mean.
❑ In Fig.2.c this assumption is denoted by the fact that
the values that u can assume lie within the same limits,
irrespective of the value of X.
11
3). No autocorrelation
❑ The error term ui is statistically independent of
one another (No error autocorrelation)
❑ This means the value which the random term
assumed in one period does not depend on the
value which it assumed in any other period.
❑ Cov (ui,uj) = 0 or cov(ɛi,ɛs|Xji,Xjs) = 0
Cov(u i u j ) = [(u i − (u i )][ u j − (u j )]
= E (u i u j ) = 0
12
4). No r/ship b/n the error and
corresponding x variate
❑ All explanatory variables are uncorrelated with
the error term.
❑ This means there is no correlation between the
random variable and the explanatory variable.
❑ If two variables are unrelated their covariance is
zero.
❑ Hence Cov (ui,xi) = 0 or cov(ɛi,Xji) = 0.
15
7). n > K+1
❑ Number of observations (n) > number of
parameters estimated (K+1).
❑ Number of parameters is K+1 in this case ( β0,
β1, …, βK )
❑ This ensures that the regression model can be
properly estimated, leading to reliable and
interpretable results.
16
8). No perfect multicollinearity
❑ This is the Additional Assumption under the
MLR
❑ The explanatory variables are not perfectly
linearly correlated.
❑ That is, no exact linear relation exists b/n any
subset of explanatory variables.
❑ In the presence of perfect (deterministic) linear
r/ship b/n or among any set of the Xjs, the
impact of a single variable ( j ) cannot be
identified.
17
Model Specification
❑ Before any equation can be estimated, the
model must be completely specified.
❑ Broadly speaking, specifying an econometric
equation consists of the following:
18
Specification error can arise in a number of
ways:
19
3.4 A Model with K Explanatory Variables
Yi = 0 + 1 X 1i + 2 X 2i + + K X Ki + i
ˆ ˆ ˆ ˆ
Y1 = 0 + 1 X 11 + 2 X 21 + + K X K 1 + e1
Y = ˆ + ˆ X + ˆ X + + ˆ X + e
2 0 1 12 2 22 K K2 2
Y3 = ˆ0 + ˆ1 X 13 + ˆ 2 X 23 + + ˆ K X K 3 + e3
Yn = ˆ0 + ˆ1 X 1n + ˆ 2 X 2 n + + ˆ K X Kn + en
• 0 is the coefficient attached to the constant term (which we called
20 before).
Y1 1 X 11 X 21 X 31 X K1 βˆ 0 e 1
Y 1 ˆ
2 X 12 X 22 X 32 X K2 β 1 e 2
Y3 = 1 X 13 X 23 X 33 X K3 • βˆ 2 + e 3
Yn 1 X 1n X 2n X 3n X Kn βˆ K e n
n ( K + 1) ( K + 1) 1
n 1 n 1
e = Y − X̂
22
3.5 A Model with Two Explanatory Variables
variables.
❑ Where
𝑦𝑖 = 𝑌𝑖 − 𝑌ത𝑖 ,
𝑥1𝑖 = 𝑋1𝑖 − 𝑋ത1𝑖 and
𝑥2𝑖 = 𝑋2𝑖 − 𝑋ത2𝑖
25
❑ The RSS = σ 𝜀𝑖 2
❑ 𝑅𝑆𝑆 = 𝑌𝑖 − 𝑌ത𝑖 = σ(𝑦𝑖 − 𝛽1 𝑥1𝑖 − 𝛽2 𝑥2𝑖 ) 2
26
Estimation of parameters of two-explanatory
variables model
❑ The model is: Y = + X + X + U …………… . . .. . (1)
0 1 1 2 2 i
29
❑ e 2
i = (
Yi − ˆ0 − ˆ1 X 1i − ˆ2 X 2i )2
. . . . . . . . . . . . . . (6)
❑
ei2 (
= −2 Yi − ˆ0 − ˆ1 X 1i − ˆ 2 X 2i = 0 ) …. . . . . . . . . . . . (7)
ˆ 0
ei2 ( )
❑
ˆ
= −2 X 1i Yi − ˆ0 − ˆ1 X 1i − ˆ2 X 1i = 0 …. . . . . . . . . . . . (8)
1
e = −2 (Y )
. .……..…….(9)
2
❑ ˆ 2
i
X 2i i − ˆ 0 − ˆ1 X 1i − ˆ 2 X 2i = 0
30
❑ Summing from 7 to 9, the multiple regression
equation produces three Normal Equations:
❑ Y = nˆ + ˆ X + ˆ X . . . . . . . . . . . . (10)
0 1 1i 2 2i
❑ 1i i 0 1i 1 1i 2 1i 1i
X Y = ˆ X + ˆ X 2 + ˆ X X
. . . . . . . . . . . . (11)
❑ 2i i 0 2i 1 1i 2i 2 2i
X Y = ˆ X + ˆ X X + ˆ X 2
. . . . . . . . . . . . (12)
31
❑ From (10) we obtain ̂ 0
1i i
X Y − nY X 1i = ˆ
2 (X 1i
2
− nX
2 ˆ
1i ) + 2 (X 1i X 2 − nX 1 X 2 ) . . . . (14)
❑ We know that
( X − Y )
i i = (X iYi − nX iYi ) = xi yi
2
(X − X i ) = (X i − nX i ) = xi
2 2 2 2
i
32
❑ Substituting the above equations in equation
(13), the normal equation (11) can be written in
deviation form as follows:
❑ 1 x y = ˆ x 2 + ˆ x x … . . . . . . .. (15)
1 1 2 1 2
33
❑ 1
x y = ˆ x 2 + ˆ x x
1 1 2 1 2
. . . . . . . . . . . . . . (17)
❑ 2
x y = ˆ x x + ˆ x 2 . . . . . . . . . . . . . . . (18)
1 1 2 2 2
x x x ˆ1 x y ………….(3
2
1 1 2 1
❑ = ..... (19)
x1 x2 x 2
2
ˆ 2 x2 y
34
❑ If we use Cramer’s rule to solve the above
matrix we obtain
❑
𝛴𝑥1 𝑦 . 𝛴𝑥22 − 𝛴𝑥2 𝑦 . 𝛴𝑥1 𝑥2 . . . . . . . . (20)
𝛽መ1 =
𝛴𝑥12 . 𝛴𝑥22 − 𝛴(𝑥1 𝑥2 )2
x2 y = X 2Y − nX 2Y
❑ We know that:
x22 = X 22 − nX 22 35
❑ We can also express ˆ1 and ˆ 2 in terms of
covariance and variances of Y , X 1 and X2
36
❑ An unbiased estimator of the variance of the
errors 𝛿 2 is given by:
❑ . . . . (24)
❑ . . . . . . . . . . . (25)
❑ .. . . . . . . . . . . (26)
37
❑ Standard error of coefficients
38
The coefficient of determination (R2): Two
explanatory variables case
❑ In the simple regression model, we introduced
R2 as a measure of the proportion of variation in
the dependent variable that is explained by
variation in the explanatory variable.
40
❑ In the present model of two explanatory variables:
not wise.
❑ An alternative measure of goodness of fit, called the
adjusted R and often symbolized as R , is usually reported
2
2
by regression programs.
43
❑ It is computed as:
ei /n−k
2
2 n −1
R = 1−
2
= 1 − (1 − R )
y / n − 1 n−k
2
. . . . . (32)
explained.
A. H 0 : 1 = 0
𝐻1 : 𝛽1 ≠ 0
B. H0 : 2 = 0
H1 : 2 0 47
❑ The null hypothesis in (A) states that, holding
variable Yi.
following tests: 48
The student’s t-test:
❑ We compute the t-ratio for each ˆ i
ˆi −
t* = ~ t n -k
SE ( ˆi )
Where
SE ( ˆ 2 )
51
3.5.2 Test of Overall Significance
❑ Throughout the previous section we were
concerned with testing the significance of the
estimated partial regression coefficients
individually.
H 0 : 1 = 2 = 3 = ............ = k = 0
used.
❑ Computed F-value
𝑅 2 /𝑘−1
or 𝐹 = … . …… (33)
ESS / k − 1 1−𝑅 2 /𝑛−𝑘
F =
RSS / n − k 54
❑ If the null hypothesis is not true, then the difference
between ESS & RSS becomes large, implying that the
constraints placed on the model by the null hypothesis
have large effect on the ability of the model to fit the
data, and the value of F tends to be large.
56
3.6 Application of Multiple Linear Regression
𝑌 = 𝑎ො + 𝛽መ1 𝑋1 + 𝛽መ2 𝑋2 + 𝑒𝑖
57
❑ On the basis of the information given below answer the
following question
X 12 = 3200 X 1 X 2 = 4300 X 2 = 400
X 22 = 7300 X 1Y = 8400 X 2Y = 13500
Y = 800 X 1 = 250 n = 25
Yi 2 = 28,000
59
❑ Since the x’s and y’s in the above formula are in
deviation form, we have to find the corresponding
deviation forms of the above given values.
❑ We know that:
𝛴𝑥1 𝑦 = 𝛴𝑋1 𝑌 − 𝑛𝑋ሜ1 𝑌ሜ x2 y = X 2Y − nX 2Y x22 = X 22 − nX 22
x1 x2 = X 1 X 2 − nX 1 X 2 x12 = X 12 − nX 12
= 4300 − (25)(10)(16) = 3200 − 25(10) 2
= 300 = 700
60
❑ Now we can compute the parameters.
𝛴𝑥1 𝑦 . 𝛴𝑥22 − 𝛴𝑥2 𝑦 . 𝛴𝑥1 𝑥2
𝛽መ1 =
𝛴𝑥12 . 𝛴𝑥22 − (𝛴𝑥1 𝑥2 )2
(400)(900)−(700)(300)
= = 0.278
(700)(900)−(300)2
(700)(700) − (400)(300)
= 2
= 0.685
700 900 − 300
61
❑ The intercept parameter can be computed using
the following formula (as equation 13).
ˆ = Y − ˆ1 X 1 − ˆ 2 X 2
= 32 − (0.278)(10) − (0.685)(16)
= 18.26
ˆ 2 x 22
B). var( ˆ ) = x
1 2
1 x 22 − (x1 x 2 ) 2
Eq. 25
ei2
̂ 2 =
n−k
❑ Where k is the number of parameters (eq. 24)
In our case k = 3 ˆ = n−e 3
2
❑ 2 i
(82.24)(900)
❑ ⇒ 𝑣𝑎𝑟( 𝛽መ1 ) = = 0.137
700 900 −(300)2
❑ መ =
𝑆𝐸(𝛽1) 𝑣𝑎𝑟( 𝛽መ1 ) = 0.137 = 0.370 (eq.27)
ˆ 2 x12
❑ var( ˆ 2 ) = (Eq.26)
x x − (x1 x2 )
2
1
2
1
2
(82.24)(700)
= 2
= 0.1067
(700)(900) − (300)
❑ =
𝐸(𝛽2) 𝑣𝑎𝑟( 𝛽መ2 ) = 0.1067 = 0.327 (eq.28)
63
C.
❑ 𝛽መ1 can be tested using students t-test
H 0 : 1 = 0
𝐻1 : 𝛽1 ≠ 0
❑ This is done by comparing the computed value
of t and critical value of t which is obtained
from the table at 2 level of significance and n-k
degree of freedom.
∗ 1
𝛽 0.278
❑ Hence; t = 1 ) = = 0.751
𝑆𝐸(𝛽 0.370
64
❑ The critical value of t from the t-table at 2 = 0.05 2 = 0.025
level of significance and 22 degree of freedom is 2.074.
❑ 𝑡𝑐 = 2.074
❑ 𝑡 ∗= 0.751
❑ ⇒ 𝑡 ∗< 𝑡𝑐
Where:
➢ tc is critical value of t and
➢ t* is calculated value of t
67
e / n − k
2
(1 − R )( n − 1)
2
❑ Adjusted R = 1 − = 1−
2 i
y / n − 1
2
n−k
(1 − 0.24)( 24)
= 1−
22
= 0.178
E.
❑ Let’s set first the joint hypothesis as
H 0 : 1 = 2 = 0 against
0.25/3 − 1
=
1 − 0.25/25 − 3
F *( 2, 22) = 3.67
69
❑ This is the computed value of F.
❑ Let’s compare this with the critical value F
at 5% level of significance and with k-1 and
n-k degrees of freedom in the numerator
and denominator respectively.
❑ F (2,22) at 5% level of significance = 3.44.
❑ F* = 3.67
❑ Fc(2,22) = 3.44
70
❑ Where F*-calculated value of F & Fc -
critical value of F
❑ F*>Fc, the decision rule is to reject H0 and
accept H1.
❑ We can say that the model is significant.
❑ I.e. the dependent variable is, at least,
linearly related to one of the explanatory
variables.
71
3.7 Dummy Variables
❑ In general, the explanatory variables in any
regression analysis are assumed to be
quantitative in nature.
75
❑ Example:
76
How do we incorporate binary information into
regression models?
❑ In the simplest case, we just add it as an independent
variable in the equation.
❑ Ex: Consider the following simple model of hourly
wage determination:
Wage = β0 + β1female + β2educ + u
❑ In the above model (equation), only two observed factors
affect wage:
➢ gender and
➢ education.
77
❑ Female = 1 when the person is female, and
female = 0 when the person is male
❑ The parameter β1 has the following
interpretation:
✓ β1 is the difference in hourly wage between
females and males, given the same amount of
education (and the same error term u).
✓ The coefficient β1 determines whether there is
discrimination against women:
80
3. Perfect Multicollinearity
❑ Perfect Multicollinearity occurs when two or
more independent variables have an
exact linear relationship between them.
❑ This problem occurs when the explanatory
variables are very highly correlated with each
other.
❑ One of the assumptions of CLRM is that there
is no exact linear relationship between the
independent variables.
❑ If two explanatory variables are significantly
related, then the OLS computer program will
find it difficult to distinguish the effects of one
variable from the effects of the other.
81
❑ The more highly correlated two (or more)
independent variables are, the more difficult it
becomes to accurately estimate the coefficients
of the true model.
❑ If two variables move identically, then there is
no hope of distinguishing between their
impacts.
❑ But if the variables are only roughly correlated,
then we still might be able to estimate the two
effects accurately enough for most purposes.
82
4. Non-normality
❑ If the error terms are not normally distributed,
inferences about the regression coefficients
(using t-tests) and the overall equation (using
the F-test) will become unreliable.
Question?
11/16/2024 AU