Deep reinforcement learning (DRL) is used in robotics for decision-making, control, and navigation tasks. It enables robots to learn from their environment by performing actions and observing the results, allowing them to adapt their behavior to achieve specific goals. DRL has been applied to various robotic applications, including grasping, object manipulation, and autonomous navigation3.
Increasing algorithm complexity in DRL leads to issues with reproducibility, as the latest algorithms require many implementation details to perform well on different levels. This complexity also necessitates detailed task design in the form of slow reward engineering. Additionally, complex RL algorithms often struggle with simple problems and can be prone to lower performance when faced with sensor noise or failure.
Simpler parametrizations in RL, such as linear functions or radial basis functions (RBF), can highlight the fragility of RL algorithms2. These simpler approaches may struggle to capture complex patterns and relationships in the data, leading to less effective policies and lower overall performance. However, they require fewer computational resources and can serve as a baseline for comparing more complex models.