Skip to content

Conversation

@vkuzo
Copy link
Contributor

@vkuzo vkuzo commented Dec 1, 2025

Summary:

Adds an e2e example of how to use torchao to quantize LLaMa 4 Scout.

Note that this needs:

Test Plan:

with-proxy time python examples/quantize_llama_4.py ~/local/tmp/20251201_test/

Reviewers:

Subscribers:

Tasks:

Tags:

[ghstack-poisoned]
@vkuzo
Copy link
Contributor Author

vkuzo commented Dec 1, 2025

Stack from ghstack (oldest at bottom):

@pytorch-bot
Copy link

pytorch-bot bot commented Dec 1, 2025

🔗 Helpful Links

🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/ao/3408

Note: Links to docs will display an error until the docs builds have been completed.

This comment was automatically generated by Dr. CI and updates every 15 minutes.

vkuzo added a commit that referenced this pull request Dec 1, 2025
Summary:

Adds an e2e example of how to use torchao to quantize LLaMa 4 Scout.

Note that this needs:
* a recent `transformers` version (higher than 4.57, not officially
  released yet so user needs to build from source)
* a recent `fbgemm_gpu` version nightly from `2025.11.22` or after
* to run this in vLLM, vllm-project/vllm#28421
  is needed (not yet landed).

Test Plan:

```bash
with-proxy time python examples/quantize_llama_4.py ~/local/tmp/20251201_test/
```

Reviewers:

Subscribers:

Tasks:

Tags:
ghstack-source-id: 3c47130
ghstack-comment-id: 3599037297
Pull-Request: #3408
@meta-cla meta-cla bot added the CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed. label Dec 1, 2025
@vkuzo vkuzo added the topic: documentation Use this tag if this PR adds or improves documentation label Dec 1, 2025
[ghstack-poisoned]
vkuzo added a commit that referenced this pull request Dec 2, 2025
Summary:

Adds an e2e example of how to use torchao to quantize LLaMa 4 Scout.

Note that this needs:
* a recent `transformers` version (higher than 4.57, not officially
  released yet so user needs to build from source)
* a recent `fbgemm_gpu` version nightly from `2025.11.22` or after
* to run this in vLLM, vllm-project/vllm#28421
  is needed (not yet landed).

Test Plan:

```bash
with-proxy time python examples/quantize_llama_4.py ~/local/tmp/20251201_test/
```

Reviewers:

Subscribers:

Tasks:

Tags:
ghstack-source-id: 76125e9
ghstack-comment-id: 3599037297
Pull-Request: #3408
@vkuzo vkuzo merged commit 5977905 into main Dec 2, 2025
35 of 51 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed. topic: documentation Use this tag if this PR adds or improves documentation

Projects

None yet

Development

Successfully merging this pull request may close these issues.

4 participants