Skip to content

Commit

Permalink
🐛 Implement add_lora in old AsyncLLMEngine
Browse files Browse the repository at this point in the history
Signed-off-by: Joe Runde <[email protected]>
  • Loading branch information
joerunde committed Jan 3, 2025
1 parent a8745c0 commit f6c940d
Showing 1 changed file with 4 additions and 0 deletions.
4 changes: 4 additions & 0 deletions vllm/engine/async_llm_engine.py
Original file line number Diff line number Diff line change
Expand Up @@ -1257,6 +1257,10 @@ async def stop_profile(self) -> None:
else:
self.engine.model_executor._run_workers("stop_profile")

async def add_lora(self, lora_request: LoRARequest) -> None:
"""Load a new LoRA adapter into the engine for future requests."""
self.engine.add_lora(lora_request)


# TODO(v1): Remove this class proxy when V1 goes default.
if envs.VLLM_USE_V1:
Expand Down

0 comments on commit f6c940d

Please sign in to comment.