We learned that positive reinforcement is better than negative reinforcement when it comes to the AI. We found that with negative reinforcement, it became possible for the AI to reach a certain point, and then stop.
The AI knew that by going forward they had a huge chance to lose all their points on the jump.
If they never made the jump, they would keep their points.
We have taught them fear, and I don't know how to feel about that.