How we suppose to submit a model that agents are connecting to?
In the submission there is an agent implementation with OpenAI client that can access LLMs. Then, for tests I used local LLM server. But how I must submit it, and how it will be launched for evaluation?
You don’t need to submit your model directly for this challenge. When you make a submission, we automatically launch an instance of each game and provide your agent with a unique MCP address for that run. Each game reports its score back to us, and we update the leaderboard accordingly.
But how you will know which model to use? Either I would like to use llama, mistral, finetuned versions or just API (openai, claude,…)