Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

About the ot_loss why optimizing the dual term's derivate instead of the OT distance ? #29

Open
Lynyanyu opened this issue May 4, 2022 · 1 comment

Comments

@Lynyanyu
Copy link

Lynyanyu commented May 4, 2022

Why setting the optimal target equals the dual term "<β*, z^>"s derivate times the prediction instead of the original OT distance? It makes sense to optimize the entire OT loss term "W(z, z^)" or its dual term "<β*, z^>" to force dot regression more sparse and accurate, but why the derivate? Is it mentioned in the paper or supplements?

@Lynyanyu Lynyanyu changed the title About the ot_loss why did you use the prediction's derivative times itself instead of the OT distance ? About the ot_loss why did you use the prediction's derivate times itself instead of the OT distance ? May 4, 2022
@Lynyanyu Lynyanyu changed the title About the ot_loss why did you use the prediction's derivate times itself instead of the OT distance ? About the ot_loss why optimizing the dual term's derivate instead of the OT distance ? May 4, 2022
@henvh
Copy link

henvh commented Jan 24, 2024

I actually had the same question. And why is the distance matrix for the ot computation defined like that?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants