Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

MultiLoRA: Democratizing LoRA for Better Multi-Task Learning #320

Open
1 task
irthomasthomas opened this issue Jan 9, 2024 · 0 comments
Open
1 task
Labels
base-model llm base models not finetuned for chat llm Large Language Models llm-evaluation Evaluating Large Language Models performance and behavior through human-written evaluation sets

Comments

@irthomasthomas
Copy link
Owner

LoRA achieves remarkable resource efficiency and comparable performance when adapting LLMs for specific tasks. Since ChatGPT demonstrated superior performance on various tasks, there has been a growing desire to adapt one model for all tasks. However, the explicit low-rank of LoRA limits the adaptation performance in complex multi-task scenarios. LoRA is dominated by a small number of top singular vectors while fine-tuning decomposes into a set of less important unitary transforms. In this paper, we propose MultiLoRA for better multi-task adaptation by reducing the dominance of top singular vectors observed in LoRA. MultiLoRA scales LoRA modules horizontally and change parameter initialization of adaptation matrices to reduce parameter dependency, thus yields more balanced unitary subspaces. We unprecedentedly construct specialized training data by mixing datasets of instruction follow, natural language understanding, world knowledge, to cover semantically and syntactically different samples. With only 2.5% of additional parameters, MultiLoRA outperforms single LoRA counterparts and fine-tuning on multiple benchmarks and model scales. Further investigation into weight update matrices of MultiLoRA exhibits reduced dependency on top singular vectors and more democratic unitary transform contributions.

Suggested labels

{ "key": "multi-task-learning", "value": "Adapting one model for multiple tasks using MultiLoRA framework for better performance" }

@irthomasthomas irthomasthomas added llm Large Language Models base-model llm base models not finetuned for chat New-Label Choose this option if the existing labels are insufficient to describe the content accurately llm-evaluation Evaluating Large Language Models performance and behavior through human-written evaluation sets and removed New-Label Choose this option if the existing labels are insufficient to describe the content accurately labels Jan 9, 2024
@irthomasthomas irthomasthomas changed the title [2311.11501] MultiLoRA: Democratizing LoRA for Better Multi-Task Learning MultiLoRA: Democratizing LoRA for Better Multi-Task Learning Mar 16, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
base-model llm base models not finetuned for chat llm Large Language Models llm-evaluation Evaluating Large Language Models performance and behavior through human-written evaluation sets
Projects
None yet
Development

No branches or pull requests

1 participant