-
Notifications
You must be signed in to change notification settings - Fork 234
Issues: pytorch/ao
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
[Bug] FSDP2 FP8 compatibility problem with nn.Linear layers (GPU count > out_features)
float8
#1938
opened Mar 24, 2025 by
HIT-cwh
Torchao's CPU overhead counteracts the performance benefit of using quantization kernel.
#1930
opened Mar 21, 2025 by
LuFinch
Accelerate activation sparsity with activation compression
good first issue
Good for newcomers
#1920
opened Mar 18, 2025 by
jcaip
Can FP8 GEMM be enabled via module hooks instead of module swapping?
#1887
opened Mar 14, 2025 by
zigzagcai
Torchao opt resuming from ckpt requires
weights_only=False
optimizer
#1885
opened Mar 13, 2025 by
felipemello1
Status of prototype features
rfc
topic: deprecation
Use this tag if this PR deprecates a feature
tracker
#1807
opened Mar 1, 2025 by
msaroufim
6 of 11 tasks
What kind of layers are optimized by torchao on a RTX 4090?
performance
question
Further information is requested
#1805
opened Mar 1, 2025 by
naiveen
mx cast to mxfp8 across dim0 and dim1 should be performant
mx
triaged
#1788
opened Feb 26, 2025 by
vkuzo
error in using torchao and torch compile on rtx 4090
autoquant
bug
Something isn't working
performance
quantize
#1775
opened Feb 25, 2025 by
zhangvia
torch.compile cast to mxfp8 with blocked scales should be performant
#1773
opened Feb 24, 2025 by
vkuzo
[QST] About NaNs generated during FP16->FP8 quantization
high priority
triage review
#1766
opened Feb 23, 2025 by
alexsamardzic
Previous Next
ProTip!
Follow long discussions with comments:>50.