Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

trtllm-bench missing support of moe_ep_size / moe_tp_size. #2577

Open
4 tasks
juewAtAmazon opened this issue Dec 16, 2024 · 0 comments
Open
4 tasks

trtllm-bench missing support of moe_ep_size / moe_tp_size. #2577

juewAtAmazon opened this issue Dec 16, 2024 · 0 comments
Labels
bug Something isn't working

Comments

@juewAtAmazon
Copy link

System Info

It appears trtllm-bench is missing support of moe_ep_size / moe_tp_size.

To evaluate MoE model's expert parallelism, e.g., in ref [1], can we get a roadmap update of trtllm-bench's support of MoE parallelisms?

Or please clarify how to get the tooling / process / scripts used in [1] below.

Thanks

1: https://developer.nvidia.com/blog/demystifying-ai-inference-deployments-for-trillion-parameter-large-language-models/

Who can help?

@ncomly-nvidia

Information

  • The official example scripts
  • My own modified scripts

Tasks

  • An officially supported task in the examples folder (such as GLUE/SQuAD, ...)
  • My own task or dataset (give details below)

Reproduction

trtllm-bench --model mistralai/Mixtral-8x22B-v0.1 build --moe_ep_size 4 --pp_size 2 --quantization FP8 --dataset /home/ubuntu/mistral-8x22b.data

Expected behavior

trtllm-bench support MoE parallelisms.

actual behavior

[TensorRT-LLM] TensorRT-LLM version: 0.15.0
Usage: trtllm-bench build [OPTIONS]
Try 'trtllm-bench build --help' for help.

Error: No such option: --moe_ep_size (Possible options: --max_batch_size, --pp_size, --tp_size)

additional notes

n/a

@juewAtAmazon juewAtAmazon added the bug Something isn't working label Dec 16, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working
Projects
None yet
Development

No branches or pull requests

1 participant