1. Introduction

General Model

We have a quantitative response Y and different predictors X1,X2,,Xp. X=(X1,,Xp)

Which we can model as

Y=f(X)+ε

Prediction

With new predictor values xo predict response Y0

Inference

Understand the relationship between predictor and response in more detail and use that to make decisions. Quantify uncertainty.

Regression Model

Regression Analysis
Given observations {(xi,yi)}i=n find relationship

yi=f(xi)+εi

where xi is assumed to be constant and E[εi]=0E[yi]=f(xi)
f describes the systematic influence of xi on yi

Prediction

Yi=f(xi)+εi,E[εi]=0

Prediction y^0=f^(x0)
To measure the accuracy of our prediction we use MSE:

MSE[f^(x0)]=E[(Y0f^(x0))2]E[(f^(x0)Y0)2]=E[(f^(x0)f(x0))2]+Var(ε0)irreducible error

With the assumption that f^(x0) is independent of ε0

E[(Y0f^(x0))2]=E[(f(x0)+ε0f^(x0))2]=E[(f(x0)f^(x0))22ε0(f(x0)f^(x0))+ε2]=E[(f(x0)f^(x0))2]+(E[ε02])+2E[ε0(f(x0)f^(x0))]=E[(f(x0)f^(x0))2]+(Var(ε0)+(E[ε0])2)+2E[ε0(f(x0)f^(x0))]=E[(f(x0)f^(x0))2]+Var(ε0)+2E[ε0]E[(f(x0)f^(x0))]by independence=E[(f(x0)f^(x0))2]+Var(ε0)

E[f^(x0)]f(x0) is called bias.

It describes how far from the truth the prediction f^(x0) is on average.  
Var(f^(x0)) describes how much variation the estimator f^(x0) has.