Summary
In this chapter, we discussed why 𝜖-greedy exploration is not the best in some cases and checked alternative modern approaches for exploration. The topic of exploration is much wider and lots of interesting methods are left uncovered, but I hope you were able to get an overall impression of the new methods and the way they should be implemented and used in your own problems.
In the next chapter, we’ll take a look at another approach to the exploration in complex enviroments: RL with human feedback (RLHF).