multi-lora batching
#14249
Replies: 1 comment 1 reply
-
vLLM support this featrue
|
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
Does vllm support batching of prompts with different lora-adapters? Is there a more detailed example code?
the example shown in <examples/offline_inference/multilora_inference.py> does not seem to express this feature
Beta Was this translation helpful? Give feedback.
All reactions