You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I tested a similar patch-based method proposed here and also a attention-based MIL on TCGA LIHC dataset. It seems that the attention-based MIL outperformed the patch-based method in AUROC by 10%. Thus I am a bit supervised by the TCGA results (patch-based) shown in the Table 3 of your paper. I guess it is because of my sub-optical settings. I would appreciate if you could share your detailed settings used for the patch-based method (epochs, batch size, optimizer, loss, etc).
Thank you very much!
The text was updated successfully, but these errors were encountered:
The performance of patch-based training (without considering MIL) majorly depends on your dataset. If your positive bags are highly unbalanced, i.e., there are a lot of negative patches in a positive slide, then the patch-based training will perform very badly, such as the experiment for Camelyon16 shown in the paper.
Hi, thank you for this fantastic work!
I tested a similar patch-based method proposed here and also a attention-based MIL on TCGA LIHC dataset. It seems that the attention-based MIL outperformed the patch-based method in AUROC by 10%. Thus I am a bit supervised by the TCGA results (patch-based) shown in the Table 3 of your paper. I guess it is because of my sub-optical settings. I would appreciate if you could share your detailed settings used for the patch-based method (epochs, batch size, optimizer, loss, etc).
Thank you very much!
The text was updated successfully, but these errors were encountered: