What I m coding: I m build an easy neural network with a weight matrix w and a second paramter u for the score. After multipling my input vector with w, the result is multiplied with a vector u to get a result as one figure and that is my score.
Where I m right now: I calculated the gradients of both two paramter with respect to my loss function.
My problem: And now i m stuck what to do next?
My solution proposal: Can I update the paramter with w = w + learingrate * w_grad (and also for u with u = u learning rate *u_grad) and do this procedure until my cost / loss value decrease... does this work? Is this correct? Is this an esay implementation of Stochastic Gradient Descent?
I m coding in Java, if you have an easy and good documented example how to optimize a neural net in an easy way, you can share it with me.
Thanks in advance!