-
Notifications
You must be signed in to change notification settings - Fork 247
Issues: pytorch/ao
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
Got unexpected low speed using quantization inference on qwen models.
#2102
opened Apr 22, 2025 by
HaoKang-Timmy
[Tracker] TorchAO activation sparsity acceleration 🚀
#2095
opened Apr 22, 2025 by
jcaip
2 of 9 tasks
Refactor torchao and tests to use model architectures from torchao.testing.model_architectures
good first issue
Good for newcomers
#2078
opened Apr 18, 2025 by
jainapurva
Dynamo error with large mesh + AdamWFp8 + bf16 stochastic rounding
bug
Something isn't working
distributed
optimizer
#2074
opened Apr 18, 2025 by
cassanof
Remove old subclass implementation to reduce maintainence cost
topic: deprecation
Use this tag if this PR deprecates a feature
#2056
opened Apr 14, 2025 by
jerryzh168
[CI] After pytorch 2.7 release, Drop support for 2.3 and 2.4 and update support to 2.5.1, 2.6 and 2.7
#2053
opened Apr 14, 2025 by
jerryzh168
[Bug] FSDP2 FP8 compatibility problem with nn.Linear layers (GPU count > out_features)
distributed
float8
#1938
opened Mar 24, 2025 by
HIT-cwh
Torchao's CPU overhead counteracts the performance benefit of using quantization kernel.
#1930
opened Mar 21, 2025 by
LuFinch
Accelerate activation sparsity with activation compression
good first issue
Good for newcomers
#1920
opened Mar 18, 2025 by
jcaip
Can FP8 GEMM be enabled via module hooks instead of module swapping?
#1887
opened Mar 14, 2025 by
zigzagcai
Previous Next
ProTip!
no:milestone will show everything without a milestone.