Professional Documents
Culture Documents
Chapter 3
At the end of this chapter
students should be able to
understand
1
Overview
3.1 Background
3.2 Multiple Linear Regression (MLR)
3.4 ANOVA
3.6 Application/Examples
2
3.1 Background
Simple regression considers the
relation between a single
independent variable and
dependent variable
Multiple regression
simultaneously considers the
influence of multiple independent
variables on a dependent variable Y
3
3.1 Background
A simple regression model
fits a regression line in 2-
dimensional space
4
3.2 Multiple Linear Regression
5
3.2 Multiple Linear Regression
Multiple linear regression (MLR), also known simply
as multiple regression, is a statistical technique that
uses several explanatory variables to predict the outcome
of a response variable. Multiple regression is an
extension of linear (OLS) regression that uses just one
explanatory variable.
6
3.2 Multiple Linear Regression
The value of the dependent variable yi is modeled as
Internal
3.2 Multiple Linear Regression
(MLR) Model
9
3.2 What is multiple
regression analysis used for?
Multiple regression analysis allows researchers to assess
the strength of the relationship between an outcome (the
dependent variable) and several predictor variables as well
as the importance of each of the predictors to the
relationship, often with the effect of other predictors
statistically eliminated.
10
3.2 Assumptions of Multiple
Linear Regression
11
3.2 Multiple Linear Regression
(MLR) – Types of Regression
12
3.3 Software Output - sample
Part 2. ANOVA
Part 1. Regression
analysis
13
3.4 Testing the significance of
Regression – t and F tests
Use of F-Test
Use of t-Test
𝐒𝐒𝐄
𝜎^ =
𝟐
𝐧− 𝑘−1
This is an unbiased estimator for s (for Population)
2
18
3.5 Model Evaluation – (ii)
Coefficient of Determination
Coefficient of determination
2 𝑆𝑆𝑇 −𝑆𝑆𝐸 𝑆𝑆𝑅 𝑆𝑆𝐸
𝑅= = =1−
𝑆𝑆𝑇 𝑆𝑆𝑇 𝑆𝑆𝑇
proportion of variability in the observed dependent variable
that is explained by the MLR model.
The coefficient of determination measures the strength of
that linear relationship, denoted by R2
The greater R2 the more successful is the MLR Model
19
3.5 Model Evaluation – (iii) The
hypothesis test of the slope (t-test)
H
𝛽 𝑖1: ≠ 0 (There is relationship between xi and y)
^
𝛽𝑖 − 𝛽𝑖 ^
𝛽𝑖 − 𝛽𝑖
𝑇= =
√
Test Statistic: T – distribution: ^2
𝜎 𝑠𝑒 ( ^
𝛽)
𝑖
𝑠 𝑠𝑥𝑥
Critical Region: |T | > tα/2, n-k-1 .
20
3.5 Model Evaluation – (iv) The
hypothesis test (p-value)
The p-value conveys information about the weight of
evidence against H0. The smaller the p–value, the greater the
evidence against H0.
When the p–value is small enough we shall reject the null
hypothesis H0 .
22
3.5 Model Evaluation – (iii) Testing the
significance of regression (F-test)
Hypotheses:
𝑀 𝑆𝑅
Test statistic: 𝐹 0=
𝑀 𝑆𝐸
where: 𝑆𝑆𝑅 𝑆𝑆𝐸
𝑀 𝑆𝑅 = , 𝑀 𝑆 𝐸=
𝑘 𝑛−𝑘 −1
Rejection criteria: 𝑀 𝑆𝑅
𝐹 0= > 𝑓 𝛼 ,𝑘,𝑛 −𝑘−1
𝑀 𝑆𝐸
23
3.5 Model Evaluation – (iii) Testing
the significance of regression (F-test)
24
3.5 Application/Examples Wire
Bond Pull Strength Data
25
Wire Bond Pull Strength Data
26
Wire Bond Pull Strength Data
Regression Statistics
Multiple R 0.990523843
R Square 0.981137483
ANOVA
df SS MS F Significance F
Regression 2 5990.771221 2995.385611 572.1672 1.07546E-19
Residual 22 115.1734828 5.235158308
Total 24 6105.944704
Upper
Coefficients Standard Error t Stat P-value Lower 95% Upper 95% Lower 95.0% 95.0%
Intercept 2.263791434 1.060066238 2.135518851 0.044099 0.065348613 4.462234256 0.06534861 4.462234
X Variable 1 2.744269643 0.093523844 29.34299438 3.91E-19 2.550313061 2.938226226 2.55031306 2.938226
X Variable 2 0.012527811 0.002798419 4.476746229 0.000188 0.006724246 0.018331377 0.00672425 0.018331
27
Wire Bond Pull Strength Data
28
Wire Bond Pull Strength Data
29
Wire Bond Pull Strength Data
Conclusion
Since > 2.074, we reject H0 ,and conclude that pull strength is
linearly related to wire length and die height
30
Wire Bond Pull Strength Data
Hypotheses:
2995.4/5.2 = 572.17
Test statistic:
31
Wire Bond Pull Strength Data
𝑀 𝑆𝑅
Rejection criteria: 𝐹 0= > 𝑓 𝛼 ,𝑘,𝑛 −𝑘−1
𝑀 𝑆𝐸
Let = 0.05. Since k = 2, n-p =22, we require to find
F(0.05,2,22).
Conclusion
ANOVA
df SS MS F Significance F
Regression 4 73.46689 18.36672 114.9671 1.56E-68
Residual 476 76.04403 0.159756
Total 480 149.5109
ANOVA
df SS MS F Significance F
Regression 3 73.38924 24.46308 153.2926 1.47E-69
Residual 477 76.12169 0.159584
Total 480 149.5109
The R square value does not change much which means that
removing Gain Density does not affect the model.
• Linear Regression:
• Nonlinear Regression: