Skip to content

Comments

Adding the support of dense models distilled from moe models with the same architecture#728

Merged
quic-rishinr merged 1 commit intoquic:mainfrom
vjanfaza:main-llama4-guard
Feb 20, 2026
Merged

Adding the support of dense models distilled from moe models with the same architecture#728
quic-rishinr merged 1 commit intoquic:mainfrom
vjanfaza:main-llama4-guard

Conversation

@vjanfaza
Copy link
Contributor

In this PR, we are adding the support of meta-llama/Llama-Guard-4-12B which is a dense model distilled form llama4 scout moe model. The changes in pytorch_transforms.py file can be applied to any dense model distilled from a moe model with supported architecture in QEfficient.

@quic-rishinr
Copy link
Contributor

@quic-amitraj please review the PR

Copy link
Contributor

@quic-amitraj quic-amitraj left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM

Signed-off-by: Vahid Janfaza <vjanfaza@qti.qualcomm.com>
@quic-rishinr quic-rishinr merged commit 4bd2239 into quic:main Feb 20, 2026
4 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants