Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

How about add MeteoRA to your survey? #3

Open
ParagonLight opened this issue Jul 27, 2024 · 2 comments
Open

How about add MeteoRA to your survey? #3

ParagonLight opened this issue Jul 27, 2024 · 2 comments

Comments

@ParagonLight
Copy link

We propose MeteoRA: Multiple-tasks Embedded LoRA for Large Language Models. Our proposed MeteoRA (Multiple-Tasks embedded LoRA) is a scalable and efficient framework that embeds multiple task-specific LoRA adapters into the base LLM via a full-mode Mixture-of-Experts (MoE) architecture. This framework also includes novel MoE forward acceleration strategies to address the efficiency challenges of traditional MoE implementations. Our evaluations, using the LlaMA2-13B and LlaMA3-8B base models equipped with 28 off-the-shelf LoRA adapters through MeteoRA, demonstrate equivalent performance with the traditional PEFT method. Moreover, the LLM equipped with MeteoRA achieves superior performance in handling composite tasks, effectively solving ten sequential problems in a single inference pass, thereby demonstrating the framework’s enhanced capability for timely adapter switching and multi-LoRA fusion.

We think this work may match your survey's focus. Please kindly consider this paper.
We also publish the code for inference and training, as well as the models published in Huggingface (MeteoRA with LlaMA2-13B and MeteoRA with LlaMA3-8B).

@withinmiaov
Copy link
Owner

Thanks for providing the information. We will check it out and try to include it in our revision.

@ParagonLight
Copy link
Author

Thank you so much!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants