-
Notifications
You must be signed in to change notification settings - Fork 53
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
internal assert failure I saw while working on Thunder #3498
Comments
@crcrpar could we get the Thunder command as well? |
@kevinstephano I added the section of repro steps to the description |
The fusion is scheduled as an inner normalization kernel with no segmentation. Looks like there's some issue with the persistent buffer logic. Here's what the fusion looks like. It's fairly complex. Fusion IR math:
I'm not sure what to blame. It seems the sync analysis is doing the right analysis since if the fusion is scheduled this way, There're some 0-dim tensors after the reduction. Not sure if they are related to the error (I suspect not). @liqiangxl Could you please take a look when you have time? |
`NVFUSER_DUMP=fusion_ir_graph` saves the dot representation of a fusion before lowering to a file named like `__tmp_fusion_ir_graph_inner_persistent_f0_c1_r0_g0.dot`. Example visualization: #3498 (comment)
I think this isn't urgent because I'm seeing some working cases under slightly different settings and my recent update leads to another issue which looks orthogonal to nvfuser. |
Steps to reproduce
scale.ndim
before applyingt
/transpose
pytorch/ao#1339 (note this requires sm >= 89)Just FYI, this script works if
nn.Linear(in_features, out_features, bias=bias)
The text was updated successfully, but these errors were encountered: