Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

MoELorA #2330

Open
moghadas76 opened this issue Jan 15, 2025 · 1 comment
Open

MoELorA #2330

moghadas76 opened this issue Jan 15, 2025 · 1 comment

Comments

@moghadas76
Copy link
Contributor

Feature request

Feature request
The paper "MoELoRA: Contrastive Learning Guided Mixture of Experts on
Parameter-Efficient Fine-Tuning for Large Language Models" introduced MoLoRA, a Mixutre-of-Experts approach using LoRA adapters. I am using it to conduct some research for my MSc thesis, and have implemented it in peft. I was wondering if this method is interesting and would be worth it to clean up my code and submit a PR.

Motivation
The motivation is to include more PEFT methods that the community can benefit from.

Your contribution
I can contribute a PR with the implementation of MoLoRA.

Motivation

Feature request
The paper "MoELoRA: Contrastive Learning Guided Mixture of Experts on
Parameter-Efficient Fine-Tuning for Large Language Models" introduced MoLoRA, a Mixutre-of-Experts approach using LoRA adapters. I am using it to conduct some research for my MSc thesis, and have implemented it in peft. I was wondering if this method is interesting and would be worth it to clean up my code and submit a PR.

Motivation
The motivation is to include more PEFT methods that the community can benefit from.

Your contribution
I can contribute a PR with the implementation of MoLoRA.

Your contribution

Feature request
The paper "MoELoRA: Contrastive Learning Guided Mixture of Experts on
Parameter-Efficient Fine-Tuning for Large Language Models" introduced MoLoRA, a Mixutre-of-Experts approach using LoRA adapters. I am using it to conduct some research for my MSc thesis, and have implemented it in peft. I was wondering if this method is interesting and would be worth it to clean up my code and submit a PR.

Motivation
The motivation is to include more PEFT methods that the community can benefit from.

Your contribution
I can contribute a PR with the implementation of MoLoRA.

@BenjaminBossan
Copy link
Member

Thank you for your proposal to add MoELoRA to PEFT. From skimming the paper, the method itself looks interesting. However, I can foresee some challenges with implementing it:

  • AFAICT, there is no reference implementation of the method.
  • Based on your name, you're also not one of the paper authors.
  • The method includes the contrastive loss component, which will make this more difficult to implement, especially so since the training code is out of scope for PEFT.

All this does not mean it could not be done, but it will most likely be challenging to implement correctly.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants