You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hi! I will update the code soon, meanwhile you can change the MODEL_PATH to meta-llama/Meta-Llama-3-70B, then launch a vllm server that host meta-llama/Meta-Llama-3-70B-Instruct by CUDA_VISIBLE_DEVICES=0,1,2,3 python -m vllm.entrypoints.openai.api_server --model meta-llama/Meta-Llama-3-70B-Instruct --tensor-parallel-size 4 --disable-log-requests --port 8000 . You can still use the chat_vllm_benchmark.py to benchmark.
馃殌 The feature, motivation and pitch
Documentation and examples are for llama2 benchmarks. We would like to run llama3 on prem benchmarks.
Alternatives
No response
Additional context
No response
The text was updated successfully, but these errors were encountered: