Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[WIP] [GenAI] Lora Finetune #7288

Open
LittleLittleCloud opened this issue Nov 6, 2024 · 1 comment
Open

[WIP] [GenAI] Lora Finetune #7288

LittleLittleCloud opened this issue Nov 6, 2024 · 1 comment
Assignees
Labels
AutoML.NET Automating various steps of the machine learning process enhancement New feature or request
Milestone

Comments

@LittleLittleCloud
Copy link
Contributor

LittleLittleCloud commented Nov 6, 2024

Lora fine-tuning is an adapter-based technique to fine-tune an LLM. It changes LLM model architecture by adding learnable lora layers to transformers. During fine-tuning, only lora weights are adjustable and the LLM weights are frozen, so it requires much less GPU memory comparing to a full-layer fine-tuning. Based on this table, it requires 16GB memory to fine-tuning a 7B size model in 16bits, which can be fit in rtx 3090, 4080 and 4090. A wider range of GPUs can be fit on 3.8B LLMs like phi-3.5-mini

API design (wip)

Package: Microsoft.ML.GenAI.Lora

interface ICausalLMLoraPipeline {} // pipeline for loading causal LM + lora layers

class LoraConfiguration // lora configuration
@michaelgsharp
Copy link
Member

@LittleLittleCloud could you add this to a Milestone for when you are planning on getting this done by?

@michaelgsharp michaelgsharp added the AutoML.NET Automating various steps of the machine learning process label Mar 10, 2025
@LittleLittleCloud LittleLittleCloud added this to the ML.NET Future milestone Mar 12, 2025
@michaelgsharp michaelgsharp removed the untriaged New issue has not been triaged label Mar 17, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
AutoML.NET Automating various steps of the machine learning process enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

2 participants