predibase / lorax

Multi-LoRA inference server that scales to 1000s of fine-tuned LLMs

Home Page:https://loraexchange.ai

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Project Roadmap

tgaddair opened this issue · comments

WIP project roadmap for LoRAX. We'll continue to update this over time.

v0.10

  • Speculative decoding adapters
  • AQLM

v0.11

  • Prefix caching
  • BERT support
  • Embedding endpoint
  • Embedding adapters
  • Classification adapters

Previous Releases

v0.9

  • Adapter memory pool

Backlog

Models

  • Llama
  • Mistral
  • GPT2
  • Qwen
  • Mixtral
  • Phi
  • Bloom
  • BERT
  • Stable-Diffusion

Adapters

Throughput / Latency

  • Paged Attention v2
  • Lookahead Decoding
  • SGMV with variable ranks
  • SGMV with tensor parallelism

Quantization

  • bitsandbytes
  • GPT-Q
  • AWQ

Usability

  • Prebuilt server wheels
  • SkyPilot usage guide
  • Example notebooks

is AWQ supported?

Hey @RileyCodes, not yet, will add that to the roadmap!

does we have tested bitsandbytes Quantization ?

Hey @abhibst, I've done some basic sanity checks on it, but haven't tested it very thoroughly. Please feel free to report any issues you encounter and I'll take a look!

Sure Thanks for confirming

How would you go about adding this in Stable Diffusion? I am really interested in experimenting with that.

Hey @sansavision, at a high level it would look a lot like the LoRA pipeline used in Diffusers: https://github.com/huggingface/api-inference-community/blob/main/docker_images/diffusers/app/pipelines/text_to_image.py#L25

A v0 shouldn't be too bad, we would basically just run a single forward pass to generate the image and perform postprocessing (as part of the existing Prefill step) and short-circuit the Decode step.

If no one has started I will start working on awq tomorrow

Nice! Thanks @flozi00, that would be awesome!

Any plans to support vision transformers from huggingface / timm? A lot of potential use cases there for deploying many classifiers. If not what would that entail? Would be open to contributing if possible.

Hey @SamGalanakis, great suggestion! The plan at the moment is to start by supporting text classifiers. Once that framework is in place for that, it should be hopefully relatively straightforward to support image classifiers as well. Happy to start a thread on Discord to discuss!

Whisper would be also very cool 😄

@tgaddair Ok clear, joined the discord will look out for it!

Hi, @tgaddair , could I know how long it will take to support the stable diffusion model?

Hey @Hap-Zhang, the plan at the moment is to add it after we add support for embedding generation and text classification. Both of those are planned for January 2024, so in the next month.

@tgaddair Okay, got it. Thank you very much for your efforts. Stay tuned for it.

If we could have OpenAI compatible endpoints that would be great too. So we can use this as drop in replacement for OpenAI models :)

Hey @AdithyanI, yes, this should be coming this week or next! See #145 to follow progress.

@tgaddair oh wow that would be awesome! Thank you so much for the work here.
If you need someone to test it out; let me know. Happy to test it out.

Is the discord still open for others to join :) ?
I followed the link of the repo, and it says it is expired.

@AdithyanI this should be landing some time today :)

#170

Hey @AdithyanI, the Discord should be available. Are you using this link?

https://discord.gg/CBgdrGnZjy

@tgaddair I asked for outlines repo authors to add support to this : outlines-dev/outlines#523
Then it would be great to have text guided generation :)

image

I don't know how hard is it to integrate that here.
Do you folks know if this is something that can be supported by LORAX?

Thanks for starting the Outlines thread @AdithyanI! Looks like the maintainer created an issue #176. Excited to explore this integration!

Would it be possible to add in context length-scaling methods like Self-Extend , Rope scaling, and/or yarn scaling? I know that llama.cpp has a good implementation of these in their server, and self-extend in particular is much more stable than rope or yarn. Having long context or doing context enhancement is super important for RAG applications.

commented

About the supported models, could you consider the ChatGLM3 ? @tgaddair

commented
  • LongLoRA

It seems that LongLoRA proposed shifted short attention is compatible with Flash-Attention, and not required during inference (ref: https://huggingface.co/Yukang/Llama-2-13b-longlora-8k#highlights), if that is true, could you share what's the planed support in LoRAX inference side? thanks @tgaddair

Do you plan on supporting AQLM to setve LoRa of Mixtral Instruct with Lorax?

Hey @thincal, the last thing we need to support LongLoRA, if I remember correctly, is #231 which @geoffreyangus is planning to pick up next week.

@remiconnesson, we have PR #233 from @flozi00 for AQLM. It's pretty close to landing, but just needs a little additional work to finish it up. If no one else picks it up, I can probably take a look in the next week or two.

Are T5 based models on the Roadmap?

@tgaddair

@remiconnesson, we have PR #233 from @flozi00 for AQLM. It's pretty close to landing, but just needs a little additional work to finish it up. If no one else picks it up, I can probably take a look in the next week or two.

Hello :) How far do you think we are for this PR to be merged? :)

Hey @remiconnesson, will probably be the next thing I take a look at after wrapping up speculative decoding this week.

@amir-in-a-cynch we can definitely add T5 to the roadmap!

Hello, will you integrate / merge / migrate to the latest hugging face text-generation-inference as it is back now with Apache 2.0 license?

Is there an expected release date for v0.11?