Questions tagged [learning-rate]

83 questions
0
votes
1 answer

Pytorch1.6 What is the actual learning rate during training?

I'd like to know the actual learning rate during training, here is my code. learning_rate = 0.001 optimizer = torch.optim.Adam(net.parameters(), lr=learning_rate) scheduler = torch.optim.lr_scheduler.MultiStepLR(optimizer, milestones=[1, 2],…
0
votes
1 answer

Darknet yolo v4 - how to get current learning rate?

When I train yolo v4 with Darknet, few times a second I get a list of variables. Is there an easy way to make Darknet binary also print learning rate? If everything fails I will make changes in…
Huxwell
  • 171
  • 2
  • 14
0
votes
1 answer

Nan losses using "Learning Rate Step Decay" Scheduler with Adam Optimizer in Keras?

I have this very deep model: def get_model2(mask_kind): decay = 0.0 inp_1 = keras.Input(shape=(64, 101, 1), name="RST_inputs") x = layers.Conv2D(256, kernel_size=(3, 3), kernel_regularizer=l2(1e-6), strides=(3, 3), padding="same")(inp_1) …
Ciccios_1518
  • 435
  • 1
  • 6
  • 13
0
votes
1 answer

Why models often benefit from reducing the learning rate during training

In Keras official documentation for ReduceLROnPlateau class (https://keras.io/api/callbacks/reduce_lr_on_plateau/) they mention that "Models often benefit from reducing the learning rate" Why is that so? It's counter-intuitive for me at least,…
21kc
  • 21
  • 6
0
votes
1 answer

Decaying the learning rate from the 100th epoch

Knowing that learning_rate = 0.0004 optimizer = torch.optim.Adam( model.parameters(), lr=learning_rate, betas=(0.5, 0.999) ) is there a way of decaying the learning rate from the 100th epoch? Is this a good practice: decayRate =…
Robert
  • 7
  • 3
0
votes
1 answer

Directly update the optimizer learning rate

I have a specific learning rate schedule in mind. It is based on the epoch but differs from the generally available ones I am aware of including StepLR. Is there something that would perform the equivalent…
WestCoastProjects
  • 58,982
  • 91
  • 316
  • 560
0
votes
1 answer

How can I converge loss to a lower value?(tensorflow)

I used tensorflow object detection API. Here is my environment. All images are from coco API Tensorflow version : 1.13.1 Tensorboard version : 1.13.1 Number of test images : 3000 Number of train images : 24000 Pre-trained model : SSD mobilenet v2…
0
votes
1 answer

Tensorboard plot ReduceLROnPlateau

I keep failing to plot my learning rate in tensorboard because I am using the ReduceLROnPlateau as following: tensorboard_callback = tf.keras.callbacks.TensorBoard(log_dir=results_path, histogram_freq=1) reduce_lr =…
Ollowain
  • 55
  • 6
0
votes
2 answers

How to change the learning rate in PyTorch (1.6)

I am using PyTorch and I want to change the learning rate after some epochs. However, the code that is provided on most documentations, which is: optimizer = torch.optim.Adam([ dict(params=model.parameters(), lr=learning_rate), ]) …
Timbus Calin
  • 13,809
  • 5
  • 41
  • 59
0
votes
1 answer

tf.compat.v1.train.exponential_decay: global step = 0

To find out how to implement both a ANN with exponential decay as well as a with a constant learning rate I looked it up here: https://www.tensorflow.org/api_docs/python/tf/compat/v1/train/exponential_decay I have some questions: ... global_step =…
0
votes
0 answers

Cant build a deep learning model as mentioned in the arxiv paper

I am trying to replicate this paper using the same dataset. But somehow they trained and achieved(80%) a lot better accuracy than I usually(65%) get. My data is imbalanced 300:500(filler words: non-filler words). The model tends to overfit in the…
0
votes
2 answers

GridSearchCV for learning rate

I'm trying to use GridSearchCV to find the best parameters in a CNN but when I try to find the best combination for learning rate and batch size the code doesn't work (it works if I use instead of learning rate, epochs). Any idea why it is not…
juanc409
  • 15
  • 1
  • 5
0
votes
1 answer

Learning rate curve going backwards for cnn

I'm working on recognizing the numbers 3 and 7 using MNIST data set. I'm using cnn_learner() function from fastai library. When I plotted the learning rate, the curve started going backwards after a certain value on X-axis. Can someone please…
0
votes
1 answer

Monitor the Learning Rate of the InverseTimeDecay() - float() argument must be a string or a number, not 'InverseTimeDecay'

My goal is to have a look at the learning rate progression of the Adam optimizer, on which I apply InverseTimeDecay schedule. So I want to check if the learning rate actually decreases. Having checked this question on stack overflow, I made similar…
NikSp
  • 1,262
  • 2
  • 19
  • 42
0
votes
0 answers

Why is learning rate get constant result (not change)?

When I use object detection with Faster R-CNN I get this result in tensorboard: Why do I get a constant result in my learning rate? What does that mean? Shouldn't my learning rate change each iteration?