Underfitting occurs when a machine learning model is too simple to capture the underlying patterns in the training data. AWS documentation explains that an underfit model performs poorly on both training and test datasets, which directly matches the scenario described.
In this case, the model shows low accuracy during training and evaluation, indicating that it has not learned sufficient relationships from the data. AWS identifies common causes of underfitting as insufficient model complexity, inadequate feature representation, overly aggressive regularization, or insufficient training time.
Underfitting is different from overfitting. Overfitting occurs when a model performs well on training data but poorly on test data, which is not the situation here. Hallucination applies to generative AI outputs, not supervised ML model accuracy. Cross-validation is a model evaluation technique, not a cause of poor performance.
AWS emphasizes the importance of diagnosing underfitting early in the model development lifecycle. Remedies include increasing model complexity, adding relevant features, reducing regularization, or selecting a more expressive algorithm. These steps allow the model to better learn from the data and improve accuracy across both training and test sets.
AWS machine learning best practices clearly associate low performance on both datasets with underfitting, making this the most likely cause of the problem described.
Submit