diplomacy_cicero

Report Abuse

Basic Information

This repository contains the research code, training scripts, checkpoints, and configuration used to build and evaluate Cicero and Diplodocus, two agents for the board game Diplomacy described in peer-reviewed papers. It includes language modeling and generation code integrated with the ParlAI framework, dialogue-free strategy model code, reinforcement learning self-play infrastructure, and example configs and agent specifications. The repo is intended to let researchers and developers reproduce experiments, run pretrained agents in self-play or in matches against baseline agents, and re-train or fine-tune models given sufficient compute. It also includes data and visualizations for consented experiment games and a separately licensed subdirectory with utilities for rendering and online connection to web play platforms.

Links

App Details

Features
Contains pre-trained model checkpoints and a download script for model files with an associated model-weights license. Language modeling and dialog code lives in a parlai_diplomacy subfolder and uses the ParlAI framework. Strategic planning logic for Cicero and Diplodocus is located in agent modules including files implementing bilateral and correlated planning and the bqre1p agent. Dialogue-free model architectures and supervised training scripts are provided along with RL self-play training code. A HeyHi config framework, Agent protobuf schemas, example configs in conf/, and run.py act as front-end for running tasks. The repo also includes build and installation instructions, tests, pre-commit hooks, and a separately licensed fairdiplomacy_external subtree for rendering and web integration.
Use Cases
This codebase helps researchers reproduce the Science and ICLR experiments by providing the full training and evaluation pipeline for language-enabled strategic agents in Diplomacy. It enables running simulations and comparisons between agents, launching agents on web platforms, and performing large-scale RL training on clusters or locally with appropriate hardware. The provided configs, protobuf schemas, and run-time framework simplify running tasks such as agent comparison, self-play, and supervised pre-training. Experiment game JSONs and visualizations support analysis of agent behavior and human-AI interactions. The repository also offers testing, formatting hooks, and documentation to support extension, benchmarking, and follow-up research on multi-agent negotiation and language-grounded strategic play.

Please fill the required fields*