-
Notifications
You must be signed in to change notification settings - Fork 216
Issues: pytorch/ao
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
int4_weight_only
api got error when saving transformers models
#1704
opened Feb 12, 2025 by
jiqing-feng
Model size after quantization
quantize
question
Further information is requested
#1701
opened Feb 11, 2025 by
TaylorYangX
[DOC] Questions on Integrating a New CPU Operator into TorchAO?
cpu
question
Further information is requested
#1699
opened Feb 11, 2025 by
Zijie-Tian
[
Fp8 Training Feature Request
] Smooth SwiGlu and Configurable AdamWFp8
#1691
opened Feb 10, 2025 by
vasqu
migration of
quantize_
workflow configuration from callables to configs
#1690
opened Feb 10, 2025 by
vkuzo
Performance comparison
NF4Tensor
vs. BNB Params4bit
performance
#1686
opened Feb 10, 2025 by
psinger
[Feature Request] Add bias support for torchao/experimental ops
#1675
opened Feb 5, 2025 by
metascroy
Tensor subclass methods for Further information is requested
DTensor
and FSDP2
question
#1664
opened Feb 5, 2025 by
jeromeku
[Needs more investigation] Something isn't working
quantize
int8_weight_only
via quantize_()
API on torch.float16
models results in NaN values across multiple CPU architectures
bug
#1662
opened Feb 4, 2025 by
vmpuri
[Doc] gemlite version
question
Further information is requested
topic: documentation
Use this tag if this PR adds or improves documentation
#1653
opened Feb 3, 2025 by
bhack
Unittests Migration Progress
good first issue
Good for newcomers
#1621
opened Jan 26, 2025 by
osbm
24 of 74 tasks
Wrong result and no speedup with SemiSparseLinear from Torchao compared to torch.nn.Linear
#1617
opened Jan 24, 2025 by
lin-ht
int4_weight_only
Slows Down torch.nn.Linear
for Llama2 7B Shapes
high priority
performance
triage review
#1606
opened Jan 23, 2025 by
mostafaelhoushi
Float8 with FSDP and DelayedScaling: 'WeightWithDelayedFloat8CastTensor' object has no attribute '_tensor'.
float8
#1605
opened Jan 23, 2025 by
fmo-mt
[float8] Add support for blockwise fp8 quantization scheme used in DeepSeek v3
float8
inference
topic: new feature
Use this tag if this PR adds a new feature
#1594
opened Jan 22, 2025 by
danielvegamyhre
Sparsity training together with FP8 linear layer in torchao
enhancement
New feature or request
sparsity
#1574
opened Jan 17, 2025 by
goldhuang
Previous Next
ProTip!
Find all open issues with in progress development work with linked:pr.