Skip to content

Repository for the submission of NeurIPS Datasets and Benchmarks Track 2022.

Notifications You must be signed in to change notification settings

SecurityGames/oef

Repository files navigation

Offline Equilibrium Finding - OEF

Repository for the submission of NeurIPS Datasets and Benchmarks Track 2022.

Dataset

The dataset in this repository includes three types of datasets for every game: random dataset, expert dataset and learning dataset. All dataset used in this work are avaiable at https://drive.google.com/drive/folders/1Y4hnkQ8hk2b81lbMaYWu26WTEF3L75FM?usp=sharing. The data entry in dataset is [current_game_state, player_id, legal_actions, action, next_game_state, next_legal_actions, next_player, reward, done, chance_node].

  • current_game_state: a list of every player's infomation state list
  • player_id: the player should take actions at curretn game state
  • legal_actions: the available action set of current game state
  • action: the selected action
  • next_game_state: a list of every player's infomation state list after excuting the action
  • next_legal_actions: the available action set of next game state
  • next_player: the player should take actions at next game state
  • reward: the rewards of every player of next game state
  • done: whether the next game state is a end state
  • chance_node: whether the next game state is a chance state

We also provide the code used to generate dataset in generate_dataset folder, therefore, the offline dataset of other games can be get using the code.

How to run the code

  • Create a virtual python environment
  • Install the requirement packages in the requirements.txt
  • Behavior Cloning Algorithm: run the train_bc_policy.py file to get the behavior cloning policy by modifying the game and corresponding dataset in that file
  • Model-based Algorithm: first run the train_env_model.py file to train the environment model by modifying the game and corresponding dataset in that file and then run OEF-CFR or OEF-PSRO algorithm to get the model-based policy based on the trained environment model
  • OEF-CFR: run the run_mb_deep_cfr.py file in the oef_cfr folder to run OEF-CFR algorithm on the trained environment model
  • OEF-CFR: run the run_mb_psro.py file in the oef_psro folder to run OEF-PSRO algorithm on the trained environment model
  • BC+MB: run the combine_bc_mb.py file to find proper weights for the behavior cloning policy and the model-based policy to get a best combination policy.

About

Repository for the submission of NeurIPS Datasets and Benchmarks Track 2022.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages