The Conference for Machine Learning Innovation

Human / AI Interaction Loop Training as a New Approach for Interactive Learning With Reinforcement-Learning Agents

Session
Join the ML Revolution!
Register until April 30:
✓ Raspberry Pi or C64 Mini for free
✓Save up to 313 €
✓10 % Team Discount
Register Now
Join the ML Revolution!
Register until April 30:
✓ Raspberry Pi or C64 Mini for free
✓Save up to 313 €
✓10 % Team Discount
Register Now
Join the ML Revolution!
Register until May 28:
✓ ML Intro Day for free
✓ Raspberry Pi or C64 Mini for free
✓ Save up to $580
Register Now
Join the ML Revolution!
Register until May 28:
✓ ML Intro Day for free
✓ Raspberry Pi or C64 Mini for free
✓ Save up to $580
Register Now
Join the ML Revolution!
Register until November 7th:
✓Save up to € 210
✓10% Team Discount
Register Now
Join the ML Revolution!
Register until November 7th:
✓Save up to € 210
✓10% Team Discount
Register Now

Human / AI interaction loop training as a new approach for interactive learning with reinforcement-learning: Reinforcement-Learning (RL) in various decision-making tasks of Machine-Learning (ML) provides effective results with an agent learning from a stand-alone reward function. However, it presents unique challenges with large amounts of environment states and action spaces, as well as in the determination of rewards. This complexity, coming from high dimensionality and continuousness of the environments considered herein, calls for a large number of learning trials to learn about the environment through RL. Imitation-Learning (IL) offers a promising solution for those challenges, using a teacher’s feedback. In IL, the learning process can take advantage of human-sourced assistance and/or control over the agent and environment. In this study, we considered a human teacher and an agent learner. The teacher takes part in the agent’s training towards dealing with the environment, tackling a specific objective, and achieving a predefined goal. Within that paradigm, however, existing IL approaches have the drawback of expecting extensive demonstration information in long-horizon problems. With this work, we propose a novel approach combining IL with different types of RL methods.

Take me to the full program of Zum vollständigen Programm von Munich Munich .

This Session belongs to the Diese Session gehört zum Programm vom MunichMunich and  und BerlinBerlin program. Take me to the program of . Hier geht es zum Programm von Singapore Singapore .

Take me to the full program of Zum vollständigen Programm von Berlin Berlin .

This Session Diese Session belongs to the gehört zum Programm von MunichMunich and  und BerlinBerlin program. Take me to the current program of . Hier geht es zum aktuellen Programm von Munich Munich , Singapore Singapore or oder Berlin Berlin .

Behind the Tracks