datasplitting
Data splitting is the practice of dividing a dataset into separate subsets for training, validation, and testing in order to assess how well a model will generalize to unseen data. The primary purpose is to provide an unbiased estimate of model performance and to prevent overfitting by separating the data used to learn from the data used to evaluate.
Common split schemes include holdout, k-fold cross-validation, stratified sampling, and time-series splits. In a holdout split,
Data splitting can be combined with preprocessing steps, but care must be taken to prevent data leakage,
Evaluation metrics vary by task, including accuracy, F1, or ROC-AUC for classification, and RMSE or MAE for