COS-PLAY: Co-Evolving LLM Decision and Skill Bank Agents for Long-Horizon Game Play
Published in arXiv preprint, 2026

Abstract
Long-horizon interactive environments demand multi-step reasoning, the chaining of multiple skills over many timesteps, and robust decision-making under delayed rewards and partial observability. Large Language Models (LLMs) offer a promising alternative as game-playing agents, but they often struggle with consistent long-horizon decision-making because they lack a mechanism to discover, retain, and reuse structured skills across episodes. We present COS-PLAY, a co-evolution framework in which an LLM decision agent retrieves skills from a learnable skill bank to guide action taking, while an agent-managed skill pipeline discovers reusable skills from the agent's unlabeled rollouts to form a skill bank. Our framework improves both the decision agent to learn better skill retrieval and action generation, while the skill-bank agent continually extracts, refines, and updates skills together with their contracts. Experiments across six game environments show that COS-PLAY with an 8B base model achieves over 25.1% average reward improvement against four frontier LLM baselines on single-player game benchmarks while remaining competitive on multi-player social reasoning games.
| Project Website | Code |
|---|---|
| Project Website | GitHub Code |
Please cite our work if you found it useful,
@inproceedings{wu2026cosplay,
title={Co-Evolving {LLM} Decision and Skill Bank Agents for Long-Horizon Game Play},
author={Wu, Xiyang and Li, Zongxia and Shi, Guangyao and Duffy, Alexander and Marques, Tyler and Olson, Matthew Lyle and Zhou, Tianyi and Manocha, Dinesh},
booktitle={Conference on Language Modeling (COLM)},
year={2026}
}
