Excited to get started on this challenge. As I was working through the Python starter notebook a couple questions that have come up for me.
The starter file generates 362 predictions for what appears to be the validation data. Shouldn’t the predictions submitted be for the test data for which there are 1473 observations or am I mistaken?
What should be the form of our predictions?
In the overview the Output is 2 columns - “ID” and “Predicted Class”, but in the starter notebook the generated predictions are 4 columns - “row_id”, “normal_diagnosis_probability”, “post_alz…probability”, “pre_alz…probability”.
Yes, you are right! The starter notebook generated predictions on the validation data. During evaluation (when you submit the notebook to AIcrowd), we use test data (which is not exposed to the participants) and re-run the notebook. The test data that is used during the evaluation will have 1473 observations.
The output file should have the row ID (row_id) and probabilities for normal, pre and post-diagnosis (normal_diagnosis_probability, pre_alzheimer_diagnosis_probability, post_alzheimer_diagnosis_probability). The sum of the values of the three diagnoses should not be greater than one, that is,
This just raised a question for me. So according to the above statement, does it mean the sum of the 3 class probabilities cannot be = 1 always?
If it’s a multiclass problem the sum of probabilities will be equal to 1, however if it is a multilabel problem, there are chances that sum of 3 class probabilities might be greater than 1 or it can also be less than 1. So which one holds right here?
@thanish: this is a multiclass classification problem, so the sum of the probabilities should be equal to 1.
The description specifies that the sum is less than equals to 1, to stay true to the implementation details of the validation strategies in place. And also to communicate that this is a probability distribution after all, and cannot sum up to more than 1.