1

I believe term ‘learning rate schedule’ is a certain solution for tuning the learning rate. But at the same time, every parameter evaluating the parameter itself can be called a hyperparameter. So can the learning rate schedule be called a hyperparameter?

matthiaw91
  • 1,585
  • 7
  • 17
Lee
  • 11
  • 1

2 Answers2

2

Hyper-parameters are essentially parameters that you have to tune manually and aren't learned.

The schedule can be considered a hyper-parameter, as well as all the scheduler's parameters (patience, decay, etc.) and even the decision to use or not a scheduler can be considered a hyperparameter.

Djib2011
  • 8,068
  • 5
  • 28
  • 39
1

Learning rate scheduling can be achieved through multiple ways e.g. 1 cycle, power, performance.

Selecting the approach is first Hyperparameter.

Then, selecting the appropriate parameter value for the approach will be another set of Hyperparameters. e.g. eta0, step size etc.

10xAI
  • 5,929
  • 2
  • 9
  • 25