Replies: 1 comment
-
|
It depends on your model and usecase. But generally SGLang is a bit faster |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
I see many benchmark scripts, and I was wondering if there are aggregated results vs VLLM for different models/input lengths/output lengths so that I dont have to rerun them all.
Beta Was this translation helpful? Give feedback.
All reactions