Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Reproducing Test AUCs #14

Open
geoffreyangus opened this issue May 19, 2020 · 2 comments
Open

Reproducing Test AUCs #14

geoffreyangus opened this issue May 19, 2020 · 2 comments

Comments

@geoffreyangus
Copy link

Hi,

Thank you all for providing this repository for public use. I am trying to reproduce the results from the paper, namely the test AUC given for the 50-bag, 10-instance experiment.

I've run the implementation in this repository with the following command:

python main.py --num_bags_train 50 --num_bags_test 1000

Doing so actually overshoots the result given in the paper, by a substantial margin (0.768 (paper) vs. 0.898 (repo)). I understand that there are differences in the repository vs. the implementation in the paper (i.e. no validation set, no early stopping). However, given that the sample count in the training bags is so small, I am not convinced that such a large difference is due to the data split, and I am printing the AUC at each step for both the train and test set, which should allow me to reason about the early stopping difference. Is there something else that I am missing? Let me know, thanks!

@max-ilse
Copy link
Collaborator

max-ilse commented May 20, 2020 via email

@binli123
Copy link

binli123 commented Jun 6, 2020

Hey, I recently committed the original dataloader mnist_bags_loader.py you would have to use that instead of dataloader.py

On Tue, May 19, 2020, 4:34 AM Geoffrey Angus @.***> wrote: Hi, Thank you all for providing this repository for public use. I am trying to reproduce the results from the paper, namely the test AUC given for the 50-bag, 10-instance experiment. I've run the implementation in this repository with the following command: python main.py --num_bags_train 50 --num_bags_test 1000 Doing so actually overshoots the result given in the paper, by a substantial margin (0.768 (paper) vs. 0.898 (repo)). I understand that there are differences in the repository vs. the implementation in the paper (i.e. no validation set, no early stopping). However, given that the sample count in the training bags is so small, I am not convinced that such a large difference is due to the data split, and I am printing the AUC at each step for both the train and test set, which should allow me to reason about the early stopping difference. Is there something else that I am missing? Let me know, thanks! — You are receiving this because you are subscribed to this thread. Reply to this email directly, view it on GitHub <#14>, or unsubscribe https://github.com/notifications/unsubscribe-auth/ACNXTIRDEWDOEMZQ27NJ3LTRSHV3JANCNFSM4NETAD2A .

Hi Max, I am also getting much higher AUC values around 0.89. I returned the value Y_prob, and used all truth labels and Y_prob values to compute ROC and AUC, was there anything I did wrong? I used the newly uploaded dataloader mnist_bags_loader.py. Thanks.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants