Skip to content

Latest commit

 

History

History
402 lines (279 loc) · 12.4 KB

models.md

File metadata and controls

402 lines (279 loc) · 12.4 KB

Models

Table of Contents

  1. Model Aliases
  2. Model Alias Values

Model Aliases

To simplify using LLMInterface.sendMessage(), you can use the following model aliases:

  • default
  • large
  • small
  • agent

If no model is passed, the system will use the default model for the LLM provider. If you'd prefer to specify your model by size instead of name, pass large or small.

Aliases can simplify working with multiple LLM providers letting you call different providers with the same model names out of the box.

const openaiResult = await LLMInterface.sendMessage("openai", "Explain the importance of low latency LLMs", { model: "small" });
const geminiResult = await LLMInterface.sendMessage("gemini", "Explain the importance of low latency LLMs", { model: "small" });

Changing the aliases is easy:

LLMInterface.setModelAlias("openai", "default", "gpt-3.5-turbo");

Model Alias Values

  • default: jamba-instruct
  • large: jamba-instruct
  • small: jamba-instruct
  • agent: jamba-instruct

  • default: Llama-2-70b
  • large: Qwen/Qwen1.5-72B-Chat
  • small: alpaca-7b
  • agent: Llama-2-70b

  • default: gpt-3.5-turbo-16k
  • large: Qwen/Qwen1.5-72B-Chat
  • small: Qwen/Qwen1.5-0.5B-Chat
  • agent: gpt-4-32k-0613

  • default: mistralai/Mixtral-8x22B-Instruct-v0.1
  • large: meta-llama/Llama-3-70b-chat-hf
  • small: mistralai/Mistral-7B-Instruct-v0.1
  • agent: mistralai/Mixtral-8x22B-Instruct-v0.1

  • default: claude-3-sonnet-20240229
  • large: claude-3-opus-20240229
  • small: claude-3-haiku-20240307
  • agent: claude-3-sonnet-20240229

  • default: @cf/meta/llama-3-8b-instruct
  • large: @hf/thebloke/llama-2-13b-chat-awq
  • small: @cf/tinyllama/tinyllama-1.1b-chat-v1.0
  • agent: @cf/meta/llama-3-8b-instruct

  • default: command-r
  • large: command-r-plus
  • small: command-light
  • agent: command-r-plus

  • default: gpt-4-turbo-2024-04-09
  • large: gpt-4o
  • small: cortext-lite
  • agent: gemini-pro

  • default: openchat/openchat-3.6-8b
  • large: nvidia/Nemotron-4-340B-Instruct
  • small: microsoft/WizardLM-2-7B
  • agent: Qwen/Qwen2-7B-Instruct

  • default: deepseek-chat
  • large: deepseek-chat
  • small: deepseek-chat
  • agent: deepseek-chat

  • default: accounts/fireworks/models/llama-v3-8b-instruct
  • large: accounts/fireworks/models/llama-v3-70b-instruct
  • small: accounts/fireworks/models/phi-3-mini-128k-instruct
  • agent: accounts/fireworks/models/llama-v3-8b-instruct

  • default: forefront/Mistral-7B-Instruct-v0.2-chatml
  • large: forefront/Mistral-7B-Instruct-v0.2-chatml
  • small: forefront/Mistral-7B-Instruct-v0.2-chatml
  • agent:

  • default: mixtral-8x7b-instruct-v0-1
  • large: meta-llama-3-70b-instruct
  • small: meta-llama-3-8b-instruct
  • agent: gemma-7b-it

  • default: gemini-1.5-flash
  • large: gemini-1.5-pro
  • small: gemini-1.5-flash
  • agent: gemini-1.5-pro

  • default: gpt-neo-20b
  • large: gpt-neo-20b
  • small: gpt-neo-125m
  • agent: gpt-j-6b

  • default: llama3-8b-8192
  • large: llama3-70b-8192
  • small: gemma-7b-it
  • agent: llama3-8b-8192

  • default: meta-llama/Meta-Llama-3-8B-Instruct
  • large: meta-llama/Meta-Llama-3-8B-Instruct
  • small: microsoft/Phi-3-mini-4k-instruct
  • agent: meta-llama/Meta-Llama-3-8B-Instruct

  • default: hive
  • large: gpt-4o
  • small: gemini-1.5-flash
  • agent: gpt-4o

  • default: meta-llama/Meta-Llama-3-8B-Instruct
  • large: meta-llama/Meta-Llama-3-8B-Instruct
  • small: microsoft/phi-2
  • agent: meta-llama/Meta-Llama-3-8B-Instruct

  • default: gpt-3.5-turbo
  • large: gpt-3.5-turbo
  • small: gpt-3.5-turbo
  • agent: openhermes

  • default: mistral-large-latest
  • large: mistral-large-latest
  • small: mistral-small-latest
  • agent: mistral-large-latest

  • default: meta-llama/Meta-Llama-3-8B-Instruct
  • large: google/gemma-2-9b-it
  • small: microsoft/Phi-3-mini-4k-instruct
  • agent: google/gemma-2-9b-it

  • default: Neets-7B
  • large: mistralai/Mixtral-8X7B-Instruct-v0.1
  • small: Neets-7B
  • agent:

  • default: meta-llama/llama-3-8b-instruct
  • large: meta-llama/llama-3-70b-instruct
  • small: meta-llama/llama-3-8b-instruct
  • agent: meta-llama/llama-3-70b-instruct

  • default: nvidia/llama3-chatqa-1.5-8b
  • large: nvidia/nemotron-4-340b-instruct
  • small: microsoft/phi-3-mini-128k-instruct
  • agent: nvidia/llama3-chatqa-1.5-8b

  • default: mistral-7b-instruct
  • large: mixtral-8x22b-instruct
  • small: mistral-7b-instruct
  • agent: mixtral-8x22b-instruct

  • default: llama3
  • large: llama3
  • small: llama3
  • agent:

  • default: gpt-3.5-turbo
  • large: gpt-4o
  • small: gpt-3.5-turbo
  • agent: gpt-4o

  • default: llama-3-sonar-large-32k-online
  • large: llama-3-sonar-large-32k-online
  • small: llama-3-sonar-small-32k-online
  • agent: llama-3-sonar-large-32k-online

  • default: reka-core
  • large: reka-core
  • small: reka-edge
  • agent: reka-core

  • default: mistralai/mistral-7b-instruct-v0.2
  • large: meta/meta-llama-3-70b-instruct
  • small: mistralai/mistral-7b-instruct-v0.2
  • agent: meta/meta-llama-3-70b-instruct

  • default: shuttle-2-turbo
  • large: shuttle-2-turbo
  • small: shuttle-2-turbo
  • agent: shuttle-2-turbo

  • default: gpt-4-turbo
  • large: llama-3-70b-chat
  • small: llama-2-7b-chat
  • agent: gpt-4-turbo

  • default: google/gemma-7b
  • large: mistralai/Mixtral-8x22B
  • small: google/gemma-2b
  • agent: Qwen/Qwen1.5-14B

  • default: ibm/granite-13b-chat-v2
  • large: meta-llama/llama-3-70b-instruct
  • small: google/flan-t5-xxl
  • agent: meta-llama/llama-3-70b-instruct

  • default: palmyra-x-002-32k
  • large: palmyra-x-002-32k
  • small: palmyra-x-002-32k
  • agent:

  • default: glm-4-airx
  • large: glm-4
  • small: glm-4-flash
  • agent: glm-4