This project was part of my Summer Internship at Clover IT Services Pvt Ltd.
I was responsible for end-to-end design and development, including frontend (React + Vite), backend (Firebase + Firestore), and AI integrations (ChatGPT, Gemini, DeepSeek, LLaMA, Mistral).
https://multi-model-llm-code-audit.vercel.app/
LLM Code Audit is an intelligent, developer-focused tool designed to generate, evaluate, and compare code outputs from multiple Large Language Models (LLMs), including ChatGPT, Gemini, DeepSeek, LLaMA, and Mistral.
Whether you're prototyping, benchmarking, or auditing AI-generated code, this platform helps you make data-driven decisions by providing detailed metrics around code complexity, maintainability, and readability.
With real-time analysis, visual dashboards, and history tracking, LLM Code Audit enables a transparent and measurable way to choose the best code output for your needs β all in a clean, user-friendly interface.
- Introduction
- Problem Statement
- Solution Overview
- Features
- Technical Architecture
- File Structure
- Detailed Page Descriptions
- Metrics Explanation
- Conclusion
- Contributions
LLM Code Audit helps developers analyze and compare code generated by multiple AI tools to identify the best-performing and most maintainable solution.
- AI-generated code is often unoptimized or poorly structured.
- Manually comparing outputs is tedious and error-prone.
- Lack of tools that provide actionable feedback and ranking across LLMs.
- Tools like SonarQube evaluate only post-development.
- No side-by-side LLM comparisons.
- No automatic quality metrics or recommendations.
LLM Code Audit provides:
- Multi-LLM integration
- AI code comparison
- 5 quality metrics
- Scoring
- One prompt β multiple AI outputs
- Side-by-side comparison
- Add user code for evaluation
- Detects code issues
- Auto-scores and recommends best output
- ACI (Complexity)
- AMR (Maintainability Risk)
- ARS (Readability)
- ADR (Dependency Risk)
- ARF (Redundancy)
- Stores prompts and outputs
- Tracks best LLM per prompt
- Allows re-analysis
- Minimal, modern UI
- Interactive visualizations
- Fully responsive
- React.js with Vite
- Modular components
- Styled using custom CSS
- Firebase Authentication
- Firestore for prompt/output storage
- ChatGPT
- Gemini
- DeepSeek
- LLaMA
- Mistral
node_modules/
public/
src/
βββ api/
βββ assets/
βββ backend/
βββ components/
β βββ AuthModal.jsx
β βββ Header.jsx
β βββ HistoryItem.jsx
β βββ QueryInput.jsx
β βββ ResultsDisplay.jsx
βββ hooks/
βββ pages/
β βββ AboutPage.jsx
β βββ AuthPage.jsx
β βββ ComparisonDetail.jsx
β βββ Dashboard.jsx
β βββ HistoryPage.jsx
β βββ HomePage.jsx
βββ styles/
βββ App.css
βββ App.jsx
βββ firebaseConfig.js
βββ firestoreService.js
βββ index.css
βββ main.jsx
.env
.gitignore
eslint.config.js
index.html
package-lock.json
package.json
vite.config.js
- Enter prompt
- Choose LLMs
- Add custom code
- Click "Generate" for outputs
- View AI-generated code
- Start code evaluation
- Scores and charts for each output
- Final recommendation
- View and revisit previous prompts
- Compare LLM performance over time
- Project vision
- Real-world applications
ACI = (0.5 Γ Cognitive Complexity) + (0.3 Γ Method Length) + (0.2 Γ Nesting Level)AMR = (0.4 Γ Code Smells per 100 LOC) + (0.4 Γ Duplication %) + (0.2 Γ % Uncommented Methods)ARS = (0.4 Γ Avg Line Length) + (0.4 Γ % Comments) + (0.2 Γ Naming Consistency)LLM Code Audit empowers developers to take full control over AI-generated code, ensuring quality, maintainability, and clarity before integrating into production environments.
By offering clear comparisons, actionable insights, and in-depth metrics, this tool bridges the gap between rapid AI code generation and robust software engineering standards.
Ready to reduce tech debt and make AI code work for you?
Try LLM Code Audit now!






