Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

fix SmoothQuantGatedMLP ffn_hidden_size bug #1712

Closed
wants to merge 3 commits into from

Conversation

michael200892458
Copy link

SmoothQuantMLP mlp_hidden_size is not equal config.intermediate_size when Qwen model。
mlp_hidden_size should be layer.mlp.ffn_hidden_size

@nv-guomingz
Copy link
Collaborator

nv-guomingz commented Jun 3, 2024

Hi @michael200892458 , thanks for contribuding tensorrt-llm.
May I know the background of this MR that we need to update generation.py? Is it a duplicate MR of #1685? If so, I suggest we keep this MR focusing on ffn_hidden_size bug only.

@nv-guomingz
Copy link
Collaborator

Hi @michael200892458 , after checking the latest main code, this issue had been fixed in main branch.

@nv-guomingz nv-guomingz closed this Jun 5, 2024
@nv-guomingz nv-guomingz added the triaged Issue has been triaged by maintainers label Jun 5, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
triaged Issue has been triaged by maintainers
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants