Overview
Direct Answer
Linear regression is a supervised learning algorithm that models the relationship between a dependent variable and one or more independent variables by fitting a straight line (or hyperplane in multiple dimensions) through observed data points. It assumes a linear relationship and estimates coefficients that minimise prediction error.
How It Works
The algorithm calculates optimal coefficients by minimising the sum of squared residuals—the differences between observed and predicted values. In simple regression, a single independent variable produces a 2D line; multiple regression uses the normal equation or gradient descent to solve for coefficients across n-dimensional space. The fitted model then makes predictions by applying the learned coefficients to new input data.
Why It Matters
Linear models are computationally efficient, interpretable, and require relatively small datasets, making them valuable for rapid prototyping and regulatory compliance in finance and healthcare. Their transparency—each coefficient's magnitude directly indicates variable importance—supports evidence-based decision-making where stakeholders must understand model behaviour rather than treat it as a black box.
Common Applications
Applications include sales forecasting based on historical trends, real estate price estimation from property features, medical outcome prediction (e.g., patient recovery time), and demand planning in supply chain operations. Financial institutions use it for credit risk assessment and cost-benefit analysis.
Key Considerations
The method assumes a genuine linear relationship; non-linear data produces poor predictions. Multicollinearity between independent variables, outliers, and heteroscedasticity (non-constant error variance) can degrade model performance and interpretability.
More in Machine Learning
Meta-Learning
Advanced MethodsLearning to learn — algorithms that improve their learning process by leveraging experience from multiple learning episodes.
DBSCAN
Unsupervised LearningDensity-Based Spatial Clustering of Applications with Noise — a clustering algorithm that finds arbitrarily shaped clusters based on density.
Dimensionality Reduction
Unsupervised LearningTechniques that reduce the number of input variables in a dataset while preserving essential information and structure.
t-SNE
Unsupervised Learningt-Distributed Stochastic Neighbour Embedding — a technique for visualising high-dimensional data in two or three dimensions.
Curriculum Learning
Advanced MethodsA training strategy that presents examples to a model in a meaningful order, typically from easy to hard.
Gradient Descent
Training TechniquesAn optimisation algorithm that iteratively adjusts parameters in the direction of steepest descent of the loss function.
Data Augmentation
Feature Engineering & SelectionTechniques that artificially increase the size and diversity of training data through transformations like rotation, flipping, and cropping.
Deep Reinforcement Learning
Reinforcement LearningCombining deep neural networks with reinforcement learning to enable agents to learn complex decision-making from raw sensory input.