I am trying to write a neural network MLP model from scratch. However, I am stuck on the derivative of softmax function. I know that the softmax function in python code is
def softmax(input_value):
input_value -= np.max(input_value)
return np.exp(input_value) / np.sum(np.exp(input_value))
However, I dont know how to write the code for the softmax derivative. Can anyone show me how to write the code in python? Thank you so much!