Skip to content

A no-code toolkit to finetune LLMs on your local GPU—just upload data, pick a task, and deploy later. Perfect for hackathons or prototyping, with automatic hardware detection and a guided React interface.

Notifications You must be signed in to change notification settings

RETR0-OS/ModelForge

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

ModelForge 🔧⚡

Finetune LLMs on your laptop’s GPU—no code, no PhD, no hassle.

logo

🚀 Features

  • GPU-Powered Finetuning: Optimized for NVIDIA GPUs (even 4GB VRAM).
  • One-Click Workflow: Upload data → Pick task → Train → Test.
  • Hardware-Aware: Auto-detects your GPU/CPU and recommends models.
  • React UI: No CLI or notebooks—just a friendly interface.

📖 Supported Tasks

  • Text-Generation: Generates answers in the form of text based on prior and fine-tuned knowledge. Ideal for use cases like customer support chatbots, story generators, social media script writers, code generators, and general-purpose chatbots.
  • Summarization: Generates summaries for long articles and texts. Ideal for use cases like news article summarization, law document summarization, and medical article summarization.
  • Extractive Question Answering: Finds the answers relevant to a query from a given context. Best for use cases like Retrieval Augmented Generation (RAG), and enterprise document search (for example, searching for information in internal documentation).

Installation

Prerequisites

  • Python 3.8+: Ensure you have Python installed.
  • NVIDIA GPU: Recommended VRAM >= 6GB.
  • CUDA: Ensure CUDA is installed and configured for your GPU.
  • Docker Desktop: Install Docker Desktop for your OS.
  • NVIDIA Container Toolkit: Follow these instructions to install the NVIDIA Container Toolkit.
  • HuggingFace Account: Create an account on Hugging Face and generate a finegrained access token.

Steps

  1. Clone the Repository:

    git clone https://RETR0-OS/ModelForge.git
     cd ModelForge
  2. Set HuggingFace API Key in environment variables:
    Linux:

    export HUGGINGFACE_TOKEN=your_huggingface_token

    Windows Powershell:

    $env:HUGGINGFACE_TOKEN="your_huggingface_token"

    Windows CMD:

    set HUGGINGFACE_TOKEN=your_huggingface_token
  3. Build and the Docker Images:

    docker-compose up --build

NOTE: This may take a while, especially the first time you run it. The images are quite large.

  1. Done!: Navigate to http://localhost:3000 in your browser and get started!

Running the Application Again in the Future

  1. Start the Docker Containers:
    docker-compose up
  2. Navigate to the UI:
    Open your browser and go to http://localhost:3000.

Stopping the Application

To stop the application and free up resources, open a new terminal and run:

  docker-compose down

📂 Dataset Format

{"input": "Enter a really long article here...", "output": "Short summary."},
{"input": "Enter the poem topic here...", "output": "Roses are red..."}

🛠 Tech Stack

  • transformers + peft (LoRA finetuning)
  • bitsandbytes (4-bit quantization)
  • React (UI)
  • FastAPI (Backend)
  • Docker (Containerization)
  • NVIDIA Container Toolkit (GPU support)
  • NGINX (Reverse proxy)

About

A no-code toolkit to finetune LLMs on your local GPU—just upload data, pick a task, and deploy later. Perfect for hackathons or prototyping, with automatic hardware detection and a guided React interface.

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 3

  •  
  •  
  •