Member-only story

The Delicate Balance: Underfitting, Overfitting, and the Bias-Variance Trade-Off

Tahera Firdose
3 min readNov 1, 2023

--

Hello dear readers! Today, we’ll delve into a topic that’s central to machine learning but is often a source of confusion for many: the concepts of underfitting, overfitting, and the bias-variance trade-off. These concepts are essential for any data scientist, ML enthusiast, or anyone looking to understand the intricacies of building a robust model.

Underfitting: The Oversimplified Model

What is it?

Underfitting occurs when your model is too simplistic, failing to capture the underlying patterns of the data.

Example: Imagine trying to predict the price of a house based solely on its age. It’s evident that many other factors (like location, size, amenities) play a crucial role, but if your model only considers age, it’s bound to miss out on these patterns.

What causes it?

  • Overly simplistic model architecture.
  • Not enough features in the data.
  • Model hasn’t been trained long enough.

How to spot it?
Poor performance on both training and validation data.

How to Combat Underfitting:

  • Increase model complexity: Consider using more complex models or adding more features.
  • Feature engineering: Create new features that could…

--

--

Tahera Firdose
Tahera Firdose

Written by Tahera Firdose

Datascience - Knowledge grows exponentially when it is shared

No responses yet