Hello all!

Thank you for your participation and enthusiasm during round 1! We are accepting submissions for round 2 for the Top 50 teams from Round 1.

### Hardware available for evaluations

The evaluations will run on the following hardware (AWS sagemaker `ml.p3.2xlarge`

instances):

Resources | |
---|---|

vCPUs | 8 |

RAM | 60 GB |

GPU | 16 GB Tesla V100 |

*Note: The training and rollouts for each environment will run on a separate node.*

### Evaluation configuration

The configuration used during evaluations is available at FAQ: Round 1 evaluations configuration.

### Environments

This round will run on six public (coinrun, bigfish, miner, plunder, starpilot, chaser) and four private environments.

### Scoring

The final score will be an average of mean normalized rewards for public environments and the private environment

Score = \frac{1}{12}*R_{coinrun} + \frac{1}{12}*R_{bigfish} + \frac{1}{12}*R_{miner} + \frac{1}{12}*R_{chaser} + \frac{1}{12}*R_{starpilot} \\ + \frac{1}{12}*R_{plunder} + \frac{1}{8}*R_{privateEnv1} + \frac{1}{8}*R_{privateEnv2} + \frac{1}{8}*R_{privateEnv3} + \frac{1}{8}*R_{privateEnv4}

R_{env} = Mean normalized reward for `env`

.