Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Dynamically serve LoRA modules #2860

Open
rikardradovac opened this issue Dec 20, 2024 · 1 comment
Open

Dynamically serve LoRA modules #2860

rikardradovac opened this issue Dec 20, 2024 · 1 comment

Comments

@rikardradovac
Copy link

Feature request

Do you plan on integrating dynamic serving of LoRA modules, so that new modules can be added / removed during runtime instead of having to restart the engine and add the new modules to the LORA_ADAPTERS env variable?

Motivation

I am training multiple LoRA modules and want to serve them ASAP through my inference endpoint, without the need for manual restarting and adding the new modules there. An example of it would be to send a request to some load_lora endpoint with an url/path to the new module to add.

Your contribution

Could open up a PR

@drbh
Copy link
Collaborator

drbh commented Jan 13, 2025

Hi @rikardradovac thank you for opening this issue, currently we are not planning to support dynamic lora loading in TGI. This is because we load all of the weights into memory at startup to ensure optimal performance.

It's possible to load many loras at startup, but TGI does not provide a way to add/remove these after startup. Might I recommend checking out Predibase's Lorax inference server https://github.com/predibase/lorax, I believe they support dynamic lora adapters and are build on top of TGI foundations.

I hope this is helpful! Thank you

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants