From my understanding, the goal of playout or simulation stage in MCTS is to obtain a result and award or punish the nodes on the path from root during backpropagation. (Please correct me if I am wrong)
My question is if I can use domain knowledge heuristic to get this result rather than actually simulating the game to the end.
The reason I am asking is that I am doing something similar to pathfinding and the goal is to find the path to the goal state (node), which means simulating the game to the end is very difficult in my case.