Learning by playing
Getting children (and adults) to tidy up after themselves can be a challenge, but we face an even greater challenge trying to get our AI agents to do the same. Success depends on the mastery of several core visuo-motor skills: approaching an object, grasping and lifting it, opening a box and putting things inside of it. To make matters more complicated, these skills must be applied in the right sequence.Control tasks, like tidying up a table or stacking objects, require an agent to determine how, when and where to coordinate the nine joints of its simulated arms and fingers to move correctly and achieve its objective. The sheer number of possible combinations of movements at any given time, along with the need to carry out a long sequence of correct actions constitute a serious exploration problemmaking this a particularly interesting area for reinforcement learning research.Techniques like reward shaping, apprenticeship learning or learning from demonstrations can help with the exploration problem. However, these methods rely on a considerable amount of knowledge about the taskthe problem of learning complex control problems from scratch with minimal prior knowledge is still an open challenge.Our new paper proposes a new learning paradigm called Scheduled Auxiliary Control (SAC-X) which seeks to overcome this exploration issue.Read More
Related Google News:
- Databricks on Google Cloud: an open integrated platform for data, analytics and machine learning February 17, 2021
- Uncovering Unknown Unknowns in Machine Learning February 11, 2021
- Can machine learning make you a better athlete? February 4, 2021
- Machine Learning for Computer Architecture February 4, 2021
- Evaluating Design Trade-offs in Visual Model-Based Reinforcement Learning February 3, 2021
- Learning to Reason Over Tables from Less Data January 29, 2021
- Using machine learning to improve road maintenance January 13, 2021
- How to automatically scale your machine learning predictions December 17, 2020