Machine LearningTraining Techniques

Underfitting

Overview

Direct Answer

Underfitting occurs when a machine learning model lacks sufficient complexity to learn the underlying patterns and relationships in the training data, resulting in consistently poor predictive performance on both training and test datasets. This typically indicates the model architecture or feature set is inadequate rather than a problem with data quality.

How It Works

An overly simplistic model—such as linear regression applied to non-linear data or a shallow neural network for complex classification tasks—cannot represent the decision boundaries or functional relationships present in the data. The model remains biased away from the true target function, causing high training loss and high test loss simultaneously, with no opportunity to improve through additional training iterations.

Why It Matters

Organisations investing in machine learning initiatives require models that generalise effectively to new data. Underfitting wastes computational resources and delays deployment timelines, whilst producing unreliable predictions that undermine business decisions in domains such as credit risk assessment, demand forecasting, and medical diagnostics.

Common Applications

Underfitting is frequently observed when applying simple baseline models to complex datasets in finance, healthcare analytics, and natural language processing. Examples include using a linear model for image classification, applying polynomial degree-1 regression to non-linear physical phenomena, or deploying shallow decision trees for high-dimensional fraud detection.

Key Considerations

Distinguishing between underfitting and other performance issues requires comparative analysis across model complexity levels and cross-validation strategies. Practitioners must balance model sophistication against interpretability requirements and computational constraints, avoiding the assumption that increased complexity always resolves performance deficits.

More in Machine Learning