-
Notifications
You must be signed in to change notification settings - Fork 4
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
SSL: fix MLP head and remove L2 normalization #141
Conversation
@ziw-liu started the draft. I'll let you complete it. Feel free to make other improvements in architecture as you go. |
the projected features saved during prediction is now *not* normalized
@ziw-liu , I get the following error:
|
@Soorya19Pradeep Fixed in 11fa65a. This is still very much work in progress. I have not trained a full model yet. |
Edit: see full analysis here. |
@ziw-liu our next steps:
|
Fix the sequence of batchnorm and linear in the last MLP layer of the contrastive encoder model.
Remove L2 normalization before computing triplet loss. This works best when also reducing the dimension of projections.
Refactor
light
module intorepresentation
andtranslation
to separate pipelining code for different tasks.Fix #139, fix #138.