Skip to content

Conversation

@JustinTong0323
Copy link
Collaborator

Motivation

From the perspective of user friendliness, it is essential to perform the check prior to the occurrence of the actual error. Additionally, we must ensure that the error message is rational.

This update introduces a new method, check_quantized_moe_compatibility, to validate the configuration of quantized Mixture of Experts (MoE) models. It ensures that the tensor parallel size and intermediate size adhere to the required divisibility conditions, improving error handling and model configuration validation.

Modifications

Accuracy Tests

Benchmarking and Profiling

Checklist

This update introduces a new method, check_quantized_moe_compatibility, to validate the configuration of quantized Mixture of Experts (MoE) models. It ensures that the tensor parallel size and intermediate size adhere to the required divisibility conditions, improving error handling and model configuration validation.

Signed-off-by: Xinyuan Tong <[email protected]>
@gemini-code-assist
Copy link
Contributor

Warning

You have reached your daily quota limit. Please wait up to 24 hours and I will start processing your requests again!

@hnyls2002 hnyls2002 merged commit a355794 into sgl-project:main Nov 19, 2025
92 of 109 checks passed
alisonshao pushed a commit that referenced this pull request Nov 19, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

Projects

None yet

Development

Successfully merging this pull request may close these issues.

4 participants