Hyperparameter
A hyperparameter is a configuration value external to a machine learning model that must be set before training and is not learned from the training data by standard optimization. Hyperparameters differ from model parameters, such as weights and biases, which are learned during training. Hyperparameters influence the learning process, including the capacity of the model, the speed of convergence, and the strength of regularization.
Examples include the learning rate used by optimization algorithms, regularization strength, the number of layers and
Hyperparameter tuning aims to identify values that maximize performance on a validation dataset. Common strategies include
The choice of hyperparameters can substantially affect model performance and should be reported to support reproducibility.