Linear Regression: Mathematical Foundations Linear Regression: Mathematical Foundations Linear regression is a fundamental statistical technique used to predict a real-valued output \( y \in \mathbb{R} \) for a given input data point \( x \in \mathbb{R}^D \). It assumes that the expected value of the target variable is a linear function of the input features: $$ \mathbb{E}[y \mid x] = w^\top x $$ 1. Dataset Representation Let the training dataset be represented by a feature matrix: $$ X \in \mathbb{R}^{N \times D} $$ where \( N \) is the number of data points and \( D \) is the number of features. The dataset can be expressed as: $$ X = [x_1, x_2, \dots, x_D] $$ Each \( x_i \) (for \( i = 1, \dots, D \)) is a column vector representing one feature across all samples. 2. Model Formulation A general polynomial form of regression can be written as: $$ y = w_0 + w_{11} x_1 + w_{12} x_1^2 + \dots + w_{21} x_2 + w_{22}...
Math intensive