I'm currently learning neural networks and have attempted to train an MLP to learn XOR using Back-propagation in Python. The network has two hidden layers (using Sigmoid Activation) and one output layer (also Sigmoid). The network (around 20,000 epochs, and with a learning rate of 0.1) outputs numbers close to the original class label:
prediction: 0.11428432952745145 original class output was: 0
prediction: 0.8230114358069576 original class output was: 1
prediction: 0.8229532575410421 original class output was: 1
prediction: 0.23349671680470516 original class output was: 0
When i plot the errors (for every epoch), my graph shows a steep decline, then a slight 'bump', i was under the impression that the errors would gradually reduce:
Would this be classed as converging? I've tried to adjust the learning rate, with no luck.
Thanks!