Underfitting

IT 위키

Underfitting is a common issue in machine learning where a model is too simple to capture the underlying patterns in the data. As a result, the model performs poorly on both training and test datasets, failing to achieve high accuracy. Underfitting occurs when the model lacks the capacity or complexity needed to represent the relationships within the data.

Causes of Underfitting[편집 | 원본 편집]

Several factors contribute to underfitting in machine learning models:

  • Over-Simplified Model: Models with too few parameters or too low complexity, such as linear regression for highly nonlinear data, may be unable to capture complex patterns.
  • Insufficient Training Time: Models, particularly neural networks, may underfit if they are not trained for enough epochs to learn the data’s patterns.
  • Inadequate Feature Representation: When important features are missing or irrelevant features are present, the model may struggle to learn.
  • High Regularization: Excessive regularization can simplify the model too much, reducing its ability to fit the data properly.

Signs of Underfitting[편집 | 원본 편집]

There are several indicators that a model might be underfitting:

  • Low Accuracy on Training and Test Data: The model performs poorly on both the training set and new data, indicating it hasn’t learned the underlying relationships.
  • High Bias: The model makes systematic errors, often resulting in predictions that deviate consistently from the target.
  • Simple Decision Boundaries: In models like decision trees, overly simplistic boundaries suggest the model hasn’t captured the complexity of the data.

Techniques to Avoid Underfitting[편집 | 원본 편집]

Various methods are available to mitigate or prevent underfitting in machine learning:

  • Increase Model Complexity: Choose a more complex model, such as moving from linear regression to polynomial regression or adding layers to a neural network.
  • Feature Engineering: Add new, relevant features or transform existing ones to provide more information for the model.
  • Reduce Regularization: Lowering regularization strength (e.g., L1 or L2 penalty) allows the model to learn more complex patterns.
  • Longer Training Duration: In neural networks, train the model for additional epochs to allow it to learn from the data.
  • Parameter Tuning: Optimize hyperparameters to increase model capacity, such as increasing tree depth in decision trees or adjusting learning rates in neural networks.

Examples of Underfitting-Prone Algorithms[편집 | 원본 편집]

Some algorithms are more likely to underfit if not properly tuned:

  • Linear Regression: Often underfits nonlinear data due to its simplicity.
  • Decision Trees with Shallow Depth: Trees with very few splits may fail to capture complex relationships.
  • Naïve Bayes: Due to its independence assumption, it may struggle with data that has dependent features.
  • k-Nearest Neighbors (kNN) with Large k: High values of k can lead to overly smooth decision boundaries, missing finer details in the data.

Consequences of Underfitting[편집 | 원본 편집]

Underfitting has several consequences for model performance and usability:

  • Poor Predictive Accuracy: The model’s low accuracy on both training and test data makes it unsuitable for practical applications.
  • High Bias: Underfitted models often exhibit high bias, meaning they systematically fail to capture the relationships in the data.
  • Lack of Generalization: An underfit model fails to generalize, providing inaccurate predictions on unseen data.

Related Concepts[편집 | 원본 편집]

Understanding underfitting requires familiarity with related concepts:

  • Overfitting: The opposite problem, where a model is too complex and learns noise or specific patterns in the training data.
  • Bias-Variance Tradeoff: The balance between bias (error due to overly simplistic models) and variance (error due to overly complex models).
  • Regularization: Techniques to control model complexity, which, if excessive, can lead to underfitting.
  • Cross-Validation: A technique for evaluating model performance on unseen data, helping detect underfitting or overfitting.

See Also[편집 | 원본 편집]