Artificial Intelligence & Machine Learning

Bias-Variance Tradeoff

Definition

In machine learning, the bias-variance tradeoff is the conflict in trying to simultaneously minimize two sources of error: bias (the error from erroneous assumptions in the learning algorithm) and variance (the error from sensitivity to small fluctuations in the training set).

Why It Matters

This is a fundamental concept in machine learning. A simple model has high bias and low variance (it underfits). A very complex model has low bias but high variance (it overfits). The goal is to find a model that has the best balance between the two for a given problem.

Contextual Example

A simple linear model might have high bias because it assumes the data is linear. A very deep decision tree might have high variance because it can perfectly fit the training data, including its noise.

Common Misunderstandings

  • High bias leads to underfitting.
  • High variance leads to overfitting.

Related Terms

Last Updated: December 17, 2025