Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

For the losses in your code #23

Open
themis0888 opened this issue Feb 13, 2019 · 1 comment
Open

For the losses in your code #23

themis0888 opened this issue Feb 13, 2019 · 1 comment

Comments

@themis0888
Copy link

Thanks for sharing your great work!
I read some questions and your answers about the losses, but I ask again to make it sure.
For the losses in your code, I see a similar type of losses such as
neg_loss = 2.0/self.alpha * torch.log(1 + torch.sum(torch.exp(self.alpha * (neg_pair - base))))
from the SemiHard loss, DistWeighted loss. As far as I know, those losses were a linear loss of the distance(in your code, similarity) according to the original papers. (If those losses are from the FaceNet and Sampling matteres...)
Are these losses your own 'Weight' loss which has an assumption that the data follows the distribution of the mixture of Gaussian?
Thanks :D

@bnu-wangxun
Copy link
Owner

No such assumption is needed. Weight loss is based on th idea of hard mining : to focus on harder samples, no matter positive or negative pairs.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants