Run the following command to clone the project to your local system:
git clone https://github.com/CivicTechFredericton/dei-decoder.git
cd dei-decoderNavigate to the project directory and install the necessary dependencies:
npm installOnce the dependencies are installed, you're ready to start working on the project.
First, run the development server:
npm run dev
# or
yarn dev
# or
pnpm dev
# or
bun devOpen http://localhost:3000 with your browser to see the result.
To learn more about Next.js, take a look at the following resources:
- Next.js Documentation - learn about Next.js features and API.
- Learn Next.js - an interactive Next.js tutorial.
Ollama is a platform that allows you to run large language models on your local machine. This guide provides instructions for installing Ollama on Windows, macOS, and Linux, along with steps to install the Llama 3.1 model.
A stable internet connection is required to download and install the software. Administrator/root access may be necessary for installation.
Open the Terminal. Install Ollama using Homebrew:
brew install ollama/tap/ollama
Verify the installation by checking the version:
ollama --version
Download the Ollama installer from the Ollama website. Run the installer and follow the on-screen instructions to complete the setup. After installation, open Command Prompt and verify it by checking the version:
ollama --version
Open your terminal. Add Ollama’s repository and install it using the following command:
curl -s https://ollama.com/ollama-linux.sh
Verify the installation by checking the version:
ollama --version
Installing the Llama Model After installing Ollama, you can easily download the Llama 3.1 model:
Open the terminal (or Command Prompt on Windows). Install the Llama model or any other model that Ollama supports, that you want to test. For example, llama3.1 can be installed with this code. Use your choice of model name.
ollama install llama3.1
Verify the model installation by listing all available models:
ollama list
Usage Once installed, you can start using the Llama 3.1 model with the following command:
ollama run llama3.1
If you encounter any issues, make sure:
You have administrator/root permissions for the installation process. The installation is being performed with a stable internet connection. For more details, please refer to the official Ollama documentation.
- public/: Contains public assets such as icons.
- README.md: The readme file for the project.
- src/: Contains the source code for the project.
-
app/: Contains application-specific code.
- api/: Contains API route handlers.
- documentgenerator/route.ts: API route for generating documents.
- llama/route.ts: API route for interacting with the Llama model.
- components/: Contains React components.
- decoder/
- docdecoder
- route.tsx: Todo a Document analyzer option
- linkdecoder
- route.tsx: Todo a URL link decoder
- textdecoder
- route.tsx: Current implementation text decoder
- docdecoder
- api/: Contains API route handlers.
-
layout.tsx: The root layout component for the application.
-
To Improve git commits and messaging, Use better-commits. This would help to structure the git commit messages.
- Clone the
devbranch. - Make pull requests to the
devbranch. - Follow the pull request template practices.