top of page
Department of Electrical Engineering
Control Robotics and Machine Learning Lab

Adaptive Methods for Continuous Control

Background

 

In robotics tasks, the actions that are performed at each timestep are complex high dimensional control commands. E.g., the amount of force (normalized to [-1,1]) to apply for each motor.

While there are algorithms capable of coping in such a setting, the complexity of the action space results in instability and longer training procedures.

 

Project Goal

 

Extend the Q estimation to better cope with continuous control tasks. Specifically, you will try to estimate the return using polynomials / adaptive discretization methods, such that finding the optimal action is possible analytically.

Project steps

 

  • Understand reinforcement learning and deep learning methodologies

  • Learn the basics of continuous control

  • Implement basic methods

  • Research and improve the method, compare to baselines

 

Required knowledge

 

  • Strong programming skills.

  • Any knowledge in DL and RL is an advantage.

  • Any knowledge in optimization is an advantage.

 

Environment

  • PyTorch (Python)

Comments and links

  • See the papers "Deep Deterministic Policy Gradients" and "Twin Delayed DDPG".

  • Potential publication if results are satisfying.

bottom of page