Hello all!
Thank you for your participation and enthusiasm during round 1! We are accepting submissions for round 2 for the Top 50 teams from Round 1.
Hardware available for evaluations
The evaluations will run on the following hardware (AWS sagemaker ml.p3.2xlarge
instances):
Resources | |
---|---|
vCPUs | 8 |
RAM | 60 GB |
GPU | 16 GB Tesla V100 |
Note: The training and rollouts for each environment will run on a separate node.
Evaluation configuration
The configuration used during evaluations is available at FAQ: Round 1 evaluations configuration.
Environments
This round will run on six public (coinrun, bigfish, miner, plunder, starpilot, chaser) and four private environments.
Scoring
The final score will be an average of mean normalized rewards for public environments and the private environment
Score = \frac{1}{12}*R_{coinrun} + \frac{1}{12}*R_{bigfish} + \frac{1}{12}*R_{miner} + \frac{1}{12}*R_{chaser} + \frac{1}{12}*R_{starpilot} \\ + \frac{1}{12}*R_{plunder} + \frac{1}{8}*R_{privateEnv1} + \frac{1}{8}*R_{privateEnv2} + \frac{1}{8}*R_{privateEnv3} + \frac{1}{8}*R_{privateEnv4}
R_{env} = Mean normalized reward for env
.