Reinforcement Learning

TNT members involved in this project:
Yannik Mahlau, M. Sc.
Mona Mirzaie, M. Sc.
Prof. Dr.-Ing. Bodo Rosenhahn
Maximilian Schier, M. Sc.
Frederik Schubert, M. Sc.

Reinforcement learning is a type of machine learning that involves an agent learning to make decisions by interacting with an environment. The agent receives feedback in the form of rewards or penalties for its actions, and its goal is to learn a policy that maximizes the cumulative reward over time.

The RL process involves:

  1. Exploration: The agent takes actions in the environment to gather information about the consequences of its decisions.
  2. Exploitation: The agent uses the gathered information to select actions that maximize the expected reward.
  3. Learning: The agent updates its policy based on the rewards or penalties received, with the goal of improving its decision-making over time.

The key insight of RL is that the agent learns through trial and error, without requiring explicit supervision or labeled data. By interacting with the environment and receiving feedback, the agent can learn to make optimal decisions in complex, uncertain situations.

At our institute, research on reinforcement learning primarily focuses on the following goals:

  • Sample efficiency: Learning from a small number of experiences, reducing the need for extensive exploration and data collection.
  • Interpretability: Understanding the decision-making process of the agent, enabling insights into its behavior and potential biases.
  • Unguided exploration: Allowing the agent to discover new strategies and solutions without relying on human-provided guidance or rewards.
  • High-level scene representation: Using representations that capture complex scenes, such as those from computer vision (CV) pipelines, to enable the agent to reason about the environment and make informed decisions.

These goals are important for developing reinforcement learning agents that can operate effectively in real-world environments, where data may be limited, and interpretability is crucial for trust and safety. Achieving these goals can lead to more robust, flexible, and autonomous agents.

Show all publications
  • Tristan Gottwald, Maximilian Schier, Bodo Rosenhahn
    Safe Resetless Reinforcement Learning: Enhancing Training Autonomy with Risk-Averse Agents
    European Conference on Computer Vision Workshops (ECCVW), October 2024
  • Yannik Mahlau, Frederik Schubert, Bodo Rosenhahn
    Mastering Zero-Shot Interactions in Cooperative and Competitive Simultaneous Games
    Proceedings of the 41st International Conference on Machine Learning (ICML), July 2024
  • Maximilian Schier, Christoph Reinders, Bodo Rosenhahn
    Learned Fourier Bases for Deep Set Feature Extractors in Automotive Reinforcement Learning
    2023 IEEE 26th International Conference on Intelligent Transportation Systems (ITSC), IEEE, September 2023