What are hyperparameters?
Hyperparameters are configuration variables external to the model that are set before the training process begins and control how the model learns.
- define how the model is trained. how long we train, the learning rate
What is dropout?
How does dropout work?
During training, dropout randomly “drops out” (sets to zero) a certain percentage of neurons in a layer at each training step. This means the dropped-out neurons do not contribute to the forward pass or backpropagation for that specific training example.
What’s the typical range of dropout?
0.2 to 0.5
What is the effect of dropout?
Overfitting
Learning Rate
- step size during the optimization process, particularly in gradient descent algorithms
- Dictates how much the model’s parameters are updated in response to the estimated error at each iteration
- Influences both the speed of convergence and final accuracy
- how fast the model parameters are allowed to change
What happens if the learning rate is too high?
What happens if the learning rate is too low?
What is a typical learning rate?
Max Depth
Regularization Strength
L1 Regularization
L2 regularization
Number of epochs
how many times the entire training dataset is passed through the learning algorithm
scale_pos_weight
batch_size
n_estimators
max tree depth
limits the depth of individual decision trees preventing them from becoming overly complex
Early stopping
What are some regularization techniques?
L1, L2, Dropout, Early stopping
What is the goal of regularization?
To prevent over-fitting
What hyerparameters should be adjusted for binary classification problems with imbalanced datasets?
eval_metric