No description
Find a file
2025-03-12 18:48:39 +01:00
agent add Int-HRL agent scripts 2025-03-12 18:46:09 +01:00
supplementary add results figure 2025-03-12 18:48:39 +01:00
dataset_utils.py add subgoal generation pipeline 2025-03-12 18:20:56 +01:00
LICENSE Initial commit 2024-02-27 16:09:59 +01:00
Preprocess_AtariHEAD.ipynb add subgoal generation pipeline 2025-03-12 18:20:56 +01:00
RAMStateLabeling.ipynb add subgoal generation pipeline 2025-03-12 18:20:56 +01:00
README.md add results figure 2025-03-12 18:48:39 +01:00
SubgoalsFromGaze.ipynb add subgoal generation pipeline 2025-03-12 18:20:56 +01:00
TrajectoryMatching.ipynb add subgoal generation pipeline 2025-03-12 18:20:56 +01:00

Int-HRL

This is the official repository for Int-HRL: Towards Intention-based Hierarchical Reinforcement Learning

Int-HRL uses eye gaze from human demonstration data on the Atari game Montezuma's Revenge to extract human player's intentions and converts them to sub-goals for Hierarchical Reinforcement Learning (HRL). For further details take a look at the corresponding paper.

Dataset

Atari-HEAD: Atari Human Eye-Tracking and Demonstration Dataset available at https://zenodo.org/record/3451402#.Y5chr-zMK3J
Atari-HEAD Montezuma's Revenge

To pre-process the Atari-HEAD data run Preprocess_AtariHEAD.ipynb, yielding the all_trials.pkl file needed for the following steps.

Sub-goal Extraction Pipeline

  1. RAM State Labeling: annotate Atari-HEAD data with room id and level information, as well as agent and skull location
  2. Subgoals From Gaze: run sub-goal proposal extraction by generating saliency maps
  3. Alignment with Trajectory: run expert trajectory to get order of subgoals

Intention-based Hierarchical RL Agent

The Int-HRL agent is based on the hierarchically guided Imitation Learning method (hg-DAgger/Q), where we adapted code from https://github.com/hoangminhle/hierarchical_IL_RL

results

Due to the novel sub-goal extraction pipeline, our agent does not require experts during training and is more than three times more sample efficient compared to hg-DAgger/Q.

To run the full agent with 12 separate low-level agents for sub-goal execution, run agent/run_experiment.py, for single agents (one low-level agent for all sub-goals) run agent/single_agent_experiment.py.

Extension to Venture and Hero

under construction

Citation

Please consider citing these paper if you use Int-HRL or parts of this repository in your research:

@article{penzkofer24_ncaa,
  author = {Penzkofer, Anna and Schaefer, Simon and Strohm, Florian and Bâce, Mihai and Leutenegger, Stefan and Bulling, Andreas},
  title = {Int-HRL: Towards Intention-based Hierarchical Reinforcement Learning},
  journal = {Neural Computing and Applications (NCAA)},
  year = {2024},
  pages = {1--7},
  doi = {10.1007/s00521-024-10596-2},
  volume = {36}
}
@inproceedings{penzkofer23_ala,
  author = {Penzkofer, Anna and Schaefer, Simon and Strohm, Florian and Bâce, Mihai and Leutenegger, Stefan and Bulling, Andreas},
  title = {Int-HRL: Towards Intention-based Hierarchical Reinforcement Learning},
  booktitle = {Proc. Adaptive and Learning Agents Workshop (ALA)},
  year = {2023},
  doi = {10.48550/arXiv.2306.11483},
  pages = {1--7}
}