You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
This request proposes to intergrate S2FT, a pure structure sparsity-based PEFT method that concurrently achieve state-of-theart fine-tuning performance, training efficiency, and inference scalability. More information about our NeurIPS paper can be found here: https://infini-ai-lab.github.io/S2FT-Page/, of which i'm the first author. Here is our code for the implementation: https://github.com/Infini-AI-Lab/S2FT.
Motivation
As far as I know, S2FT is the first one to offer efficient and flexible sparsity-based PEFT for LLMs (previously only some add sparsity to LoRA or use layerwise freezing). Here, we'd like to mention several importance features of S2FT:
Model Versatility: The design of our structure sparsity is based on the coupled structure in LLMs, which commonly exists in LLMs, VLMs, CNNs, and GNNs. Therefore, our method should work for many different structures.
Generalization Ability: When evaluated on more recent models such as LLaMA-3-8B, we observe that our method can outperform both LoRA and Full FT, which is because we only modified a small fraction of the original parameters. Therefore, we can maintain most advanced abilities during pre-training.
Training Efficiency: Instead of focusing on the parameter efficiency, S2FT can provide practical acceleration for model training. In our experiments, we show that S2FT can surpass LoRA in both training memory and time by 10%, which is important for resource-limited settings.
Scalable Serving: Finally, S2FT also shows good serving ability in comparison with LoRA, where we consider adapter fusion, switch, and parallelism. For these settings, S2FT always outperforms LoRA in both efficiency and performance.
Controllability: The model parameters to be updated in S2FT can be selected with user-specific functions, where LoRA cannot achieve this.
Based on these information, although S2FT is just released, we think it is new kind of PEFT method showing very good potential. And the integration of it should be benefit for future sparsity-based PEFT methods.
Your contribution
I will try to write most code for this new PEFT method based on the current PEFT
The text was updated successfully, but these errors were encountered:
Thank you for presenting this novel PEFT technique. I skimmed the paper and code and this could indeed be a nice addition to PEFT. Feel free to open a PR with your contribution. As a tip:
You can start with a draft PR with only the core implementation to get quick feedback. We can then iterate on adding tests, docs, examples.
Feature request
This request proposes to intergrate S2FT, a pure structure sparsity-based PEFT method that concurrently achieve state-of-theart fine-tuning performance, training efficiency, and inference scalability. More information about our NeurIPS paper can be found here: https://infini-ai-lab.github.io/S2FT-Page/, of which i'm the first author. Here is our code for the implementation: https://github.com/Infini-AI-Lab/S2FT.
Motivation
As far as I know, S2FT is the first one to offer efficient and flexible sparsity-based PEFT for LLMs (previously only some add sparsity to LoRA or use layerwise freezing). Here, we'd like to mention several importance features of S2FT:
Model Versatility: The design of our structure sparsity is based on the coupled structure in LLMs, which commonly exists in LLMs, VLMs, CNNs, and GNNs. Therefore, our method should work for many different structures.
Generalization Ability: When evaluated on more recent models such as LLaMA-3-8B, we observe that our method can outperform both LoRA and Full FT, which is because we only modified a small fraction of the original parameters. Therefore, we can maintain most advanced abilities during pre-training.
Based on these information, although S2FT is just released, we think it is new kind of PEFT method showing very good potential. And the integration of it should be benefit for future sparsity-based PEFT methods.
Your contribution
I will try to write most code for this new PEFT method based on the current PEFT
The text was updated successfully, but these errors were encountered: