Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Loss function #60

Open
cw314 opened this issue Sep 3, 2024 · 3 comments
Open

Loss function #60

cw314 opened this issue Sep 3, 2024 · 3 comments

Comments

@cw314
Copy link

cw314 commented Sep 3, 2024

Why does a loss function I trained increase?
image

@zw-92
Copy link

zw-92 commented Sep 27, 2024

您好,我也遇到了和您类似的问题,损失一直降不下来,并且评估指标很差。可以交流下吗。

@cw314
Copy link
Author

cw314 commented Sep 30, 2024 via email

@guipotje
Copy link
Collaborator

guipotje commented Oct 1, 2024

Hello @zw-92, @cw314, thank you for bringing this issue!
This is indeed strange at first sight, but it is the default behavior for this specific loss. Please see #35

Quoting my answer from the other issue:
I experienced this during my training as well. After several empirical tests, I concluded that the reliability loss is easier to optimize when the descriptors are random (i.e., when the network is initialized with random weights). However, as training progresses and the descriptors become non-random, the network must learn to identify 'reliable' descriptors in the embedding space.

Basically, the network quickly minimizes the loss in the beginning because descriptors are random, but when they converge, it becomes more difficult to infer if they are reliable.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants