Replies: 3 comments 4 replies
-
That's an excellent suggestion! I believe we were planning on putting a seed for torch and numpy, but I'll tag @alexey-gruzdev @psfoley @msheller for clarification. |
Beta Was this translation helpful? Give feedback.
-
Dear Organizers, |
Beta Was this translation helpful? Give feedback.
-
Dear Organizers, You may consider adding these lines to the internal codes. I think that it is also good for comparing team results at the end of the challenge. But I am not sure that they are enough :) torch.manual_seed(torch_manual_seed)
torch.cuda.manual_seed_all(torch_manual_seed)
torch.backends.cudnn.deterministic = True
torch.backends.cudnn.benchmark = False (maybe random_state can be fixed for train_test_split ) Best, |
Beta Was this translation helpful? Give feedback.
-
Dear organizers,
We compared different methods for the challenge, we have realized that the same conditions result in different scores. For example, even for the first round first collaborator training aggregation metrics are different (significant difference).
Do you plan to use seeds to simulate the same experimental conditions for different submissions? (for train-validation data split, weight initialization, etc) Or do you plan to run the experiments many times and take the average? Do you have any suggestions for us for a fair comparison?
By the way, we already fixed np.random and random seeds but it seems that they are not enough. Do you fix the seeds for the PyTorch and Cuda codes on internal codes?
In summary, what is the best way to compare different methods objectively? If we handle it by only fixing seeds, it would be great due to the runtimes.
Thank you,
Ece
Beta Was this translation helpful? Give feedback.
All reactions