Section 32 MLR Multiple Variables: lm implementation

32.1 Statistical Model

32.1.1 Null Model

\[ \large y_{i} = \beta_0 + \epsilon_{i} \]


32.1.2 Regression Model

\[ \large y_{i} = \beta_0 + \beta_1 x_{1i} + \beta_2 x_{2i} + ... + \beta_p x_{pi} + \epsilon_{i} \]


32.2 Syntax

32.2.1 Null Model

\[ \large fm \leftarrow lm(SBP \sim 1, \space data=BP) \]


32.2.2 Regression Model

\[ \large fm \leftarrow lm(SBP \sim BMI + Age + Income + DM + Ethnic, \space data=BP) \]


32.3 Assumptions

  • \(y\) is related to \(x\) by the simple linear regression model:

\[ \large y_{i} = \beta_0 + \beta_1 x_{1i} + \beta_2 x_{2i} + ... + \beta_p x_{pi} + \epsilon_{i}, \space i=1,...,n\] \[ \large E(y | X_1=x_{1i}, X_2=x_{2i}) = \hat\beta_0 + \hat\beta_1x_{1i} + \hat\beta_2x_{2i} + ... + \hat\beta_px_{pi} \]

  • The errors \(\epsilon_1, \epsilon_2, ..., \epsilon_n\) are independent of each other.

  • The errors \(\epsilon_1, \epsilon_2, ..., \epsilon_n\) have a common variance \(\sigma^2\).

  • The errors are normally distributed with a mean of 0 and variance \(\sigma^2\), that is:

\[ \large \epsilon \sim N(0,\sigma^2) \]


32.4 Hypothesis

Intercept

\[ \large H_O: \beta_0 = 0 \] \[ \large H_O: \beta_0 \ne 0 \]


Regression coefficients for k-th predictors

\[ \large H_O: \beta_k = 0 \]

\[ \large H_O: \beta_k \ne 0 \]


32.5 Investigating fitted lm object

\[ \large anova(fm) \]

\[ \large summary(fm) \]