I am doing my final year project on Machine Learning for checkers Computer game.
In this game i have automated one player ( which moves randomly ) and i want second player to learn against the randomness and become smart by more games and trials.
As i said that first player is automated so it's working quite fine, but talking about the second player , i am having some problem with it's moves.
I am using target function as
v(b) = w0+w1x1+w2x2+w3x3+w4x4+w5x5+w6x6
where x1=number of white piece x2=number of black piece x3=number of white king x4=number of black king x5=number of white pieces that are threatened x6=number of black pieces that are threatened
and w0 to w6 are weights which are to be learned by the algorithm.
Now if i take initial board state and put weights random. by putting random weights (12,-15,6,19,-5,3) we get v(b) = -36 as initially x1= 12 x2= 12 x3= 0 x4= 0 x5= 0 x6= 0
so v(b) = -36
- But -36 dosen't give me a valid position to move and learn further.
How to get a predictive value to be moved?
It would be a great help if you can contribute to my problem and put efforts to solve it.