could anybody explain me difference between calculating derivative in RBM with -h_j * x_k and - h_j(x) * x_k? I found source codes with both implementations and I am not sure which one is better (and why?)
Asked
Active
Viewed 76 times
1 Answers
0
By h_j and h_j(x) are you referring to the j^th hidden neuron sample or the j^th hidden neuron sigmoidal activation? Assuming so, if you calculate the derivative of the negative log-likelihood function, it turns out to be P(h_j = 1) * x_k.

Rudra Murthy
- 710
- 2
- 8
- 20