📢 News!
May 28 — We are excited to announce that we received $1000 from IEEE CIS for competition fund!
May 25 — Agent submission deadline has been extended to June 15.
Jan 31 — We are hosting the 2nd iteration of Dice Adventure Competition at Conference on Games 2025!
Join us on Slack channel for most up-to-date information.

Submit an Agent 🤖 Play the Game 🌐
Design, develop and train AI agents that can play the game. Play the game at your pace or attend multiple matchmaking event.

Welcome to the second Dice Adventure Human-AI Teaming Competition! We ran the first iteration of the competition in 2024 as one of the game-AI competitions hosted at 2024 Conference on Games and we are returning this year. In this competition, we are using an in-house developed game called Dice Adventure as the environment and encourage participants to explore human-AI teaming effects. Dice Adventure is a multi-player turn-based dungeon crawling game that takes three players to play. Detailed information on game rules, game tutorial and frequently asked questions can be found at Game page.

We offer two tracks and welcome participantas at all levels. To participate, you must sign up for one or both track(s). If you are interested in developing an agent, please check out the guidelines in Submit AI page for agent development track. The starter code and training environment can be accessed at the Dice-Adventure-Agents repo on GitHub. If you do not wish to submit an agent but are still interested in this competition, please check out the details on the Play page for signing up as a player in the player track. We will be hosting a few virtual and in-person matchmaking events to bring players together.

The following video provides an overview to the competition tracks, rules, short game tutorial and information on setting up the training environment. We hope to bring exciting experiences to the participants and exploring critical human-AI teaming questions to gain a better understanding in team communication, coordination, and adaptation.

Key Dates

Key Dates
All deadlines are 11:59pm UTC-12:00 (anywhere on Earth)
 
Release competition tracks and rules January 1, 2025
Release evaluation criteria January 1, 2025
Release leaderboard design January 1, 2025
Release game tutorial January 15, 2025
Release game play demo videos Januray 15, 2025
Release training environment set up tutorial January 15, 2025
Release a local build of the game March 1, 2025
Release the Unity project for the game March 1, 2025
Release agent submission template March 1, 2025
Relase competition registration links March 1, 2025
Agent submission portal opens March 1, 2025
CoG auxiliary paper deadline May 30, 2025
Agent submission portal closes May 31, 2025 June 15, 2025
Online matchmaking portal opens June 1, 2025 July 1, 2025
Online matchmaking portal closes June 30, 2025 July 30, 2025
Release competition results and announce winners July 31, 2025 August 10, 2025
Release competition report and educational materials August 20, 2025
Conference on Games August 26-29, 2025

Evaluation Criteria

As the goal of the competition is to better understand human-AI teaming dynamics, the submitted agents will not be evaluated by running against benchmarks or through bot fight. All the agents will be selected at least once and play with players in the player track as a team. Below are the scoring function showing how a team score is calculated. The score consists of a base score with a decaying factor and bonuses for each character. For each level, each character gets a bonous of 10 and they lose the bonus proportional to losing the health. If a character lose all the health in a level, it will not be eligible for the bonus.

  • d: decaying factor, we are using d=0.95 for competitions
  • i: the current level the team is at
  • j=(1, 2, 3) representing each of the three characters
  • r: the number of rounds the team took to clear a level
  • rs: if a character lose all lives (rs=0) in a level
  • h: the health a character remains
  • h_max: the maximum health a character has
  • n: the total number of levels the team has cleard

Prizes

Our competition is sponsored by the IEEE Computational Intelligence Society Education Competition Fund with a total cash prize of $1,000 USD, which will be awarded to participating teams ranking at the top three places on the leaderboard when the competition ends. The prize will be equally distributed among team members — human players and agent developers of agents. Below is the distribution of prizes:

🥇 1st place - $500 USD
🥈 2nd place - $300 USD
🥉 3rd place - $200 USD

Player track signup form: https://gatech.co1.qualtrics.com/jfe/form/SV_1IlcmOGhbH43OMm
Agent submission form: https://gatech.co1.qualtrics.com/jfe/form/SV_6Qd51ZHWarDyzu6
Agent file submission link: https://www.dropbox.com/request/5Fnu21FIHgVk9pnTmxId

Organizers

Qiao Zhang, qzhang490@gatech.edu
Glen Smith, glensmith@gatech.edu
Varun Girdhar, vgirdhar@andrew.cmu.edu
Ziyu Li, ziyul@andrew.cmu.edu
Avery Gong, xgong80@gatech.edu
Shreyas Ravishanker, sravishanker3@gatech.edu
Erik Harpstead, harpstead@cmu.edu
Christopher J. MacLellan, cmaclellan3@gatech.edu

Last updated on June 4, 2025.