Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

nan loss for shallow networks #2670

Open
numisveinsson opened this issue Jan 18, 2025 · 0 comments
Open

nan loss for shallow networks #2670

numisveinsson opened this issue Jan 18, 2025 · 0 comments
Assignees

Comments

@numisveinsson
Copy link

When deep supervision is toggled on and you have two layer networks (for smaller patched datasets), the deep supervision weights end up being np.array([0]). When they get normalized, they become NaN which end up in NaN loss values. Should be easy to fix here.

Thank you!

Numi Sveinsson
numisveinsson.com

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants