We present ZeroSumEval, an extensible framework for scaling LLM evaluation through inter-model competition. The framework enables systematic comparison of language models through competitive evaluation protocols.