I have implemented VGG-16 in tensorflow, VGG-16 is reasonably deep network, so the loss should definitely reduce. But in my code it's not reducing. But when i run the model on same batch again and again then the loss is reducing. Any idea, why such thing can happen.
VGG-net is followed from here.
Training was done on, dog-vs-cat dataset, with image size 224x224x3.
Network parameters are folloing:
lr_rate: 0.001 batch_size = 16
Find code @ GitHubGist
Output is as below: