Desik Rengarajan

Date:

Speaker

Desik Rengarajan is a PhD candidate in the Electrical and Computer Engineering Department at Texas A&M University,working on reinforcement learning. Over the course of his doctoral studies, he has developed algorithms and theories in various fields of reinforcement learning, including multi-armed bandits, multi-agent RL, contextual bandits, online-offline RL, meta RL, and has recently-explored the field of federated RL. He is currently on the job market and welcomes potential opportunities.

Website: https://sites.google.com/view/desik-rengarajan/ Google Scholar: https://scholar.google.com/citations?user=ygOY_E4AAAAJ

Abstract

In reinforcement learning, reward functions serve as an indirect method of defining the goal of the algorithm. Designing a reward function that accurately captures the task at hand while effectively guiding the learning process can be a difficult challenge, requiring expert domain knowledge and manual fine-tuning. To overcome this, it is often easier to rely on sparse rewards that merely indicate partial or complete task completion. However, this leads to RL algorithms failing to learn an optimal policy in a timely manner due to the lack of fine-grained feedback. During this talk, I will delve into the impact of sparse rewards on reinforcement and meta reinforcement learning and present algorithms that leverage sub-optimal demonstration data to overcome these challenges.

Papers: