Skip to content

Commit 870781d

Browse files
committed
New Llama4 models - offered by Groq and Fireworks + examples
1 parent 1867333 commit 870781d

File tree

4 files changed

+12
-6
lines changed

4 files changed

+12
-6
lines changed

openai-core/src/main/scala/io/cequence/openaiscala/domain/NonOpenAIModelId.scala

Lines changed: 7 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -32,6 +32,11 @@ object NonOpenAIModelId {
3232
val amazon_nova_micro_v1_0 = "amazon.nova-micro-v1:0"
3333

3434
// Llama
35+
// 400B params, 1 mil context
36+
val llama4_maverick_instruct_basic = "llama4-maverick-instruct-basic" // Fireworks AI
37+
// 107B params, 128k context
38+
val llama4_scout_instruct_basic = "llama4-scout-instruct-basic" // Fireworks AI
39+
val llama_4_scout_17b_16e_instruct = "meta-llama/llama-4-scout-17b-16e-instruct" // Groq
3540
val llama_3_3_70b_versatile = "llama-3.3-70b-versatile" // Groq
3641
val llama_3_3_70b_specdec = "llama-3.3-70b-specdec" // Groq
3742
val llama_v3p3_70b_instruct = "llama-v3p3-70b-instruct" // Fireworks AI
@@ -103,6 +108,7 @@ object NonOpenAIModelId {
103108
val meta_llama_3_8b_instruct_to_ai = "meta-llama/Meta-Llama-3-8B-Instruct" // Together AI
104109

105110
// Mistral
111+
val mistral_saba_24b = "mistral-saba-24b" // Groq
106112
// currently points to mistral-large-2407. mistral-large-2402 will be deprecated shortly.
107113
val mistral_large_latest = "mistral-large-latest" // Mistral
108114
val mistral_large_2407 = "mistral-large-2407" // Mistral
@@ -282,6 +288,7 @@ object NonOpenAIModelId {
282288
val deepseek_reasoner = "deepseek-reasoner" // Deepseek
283289
val deepseek_r1_distill_llama_70b =
284290
"deepseek-r1-distill-llama-70b" // Groq, Cerebras and Fireworks
291+
val deepseek_r1_distill_qwen_32b = "deepseek-r1-distill-qwen-32b" // Groq
285292
val deepseek_ai_deepseek_r1_distill_llama_70b_free =
286293
"deepseek-ai/DeepSeek-R1-Distill-Llama-70B-free" // Together AI
287294
val deepseek_ai_deepseek_r1_distill_llama_70b =

openai-examples/src/main/scala/io/cequence/openaiscala/examples/fireworksai/FireworksAICreateChatCompletion.scala

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -25,7 +25,7 @@ object FireworksAICreateChatCompletion extends ExampleBase[OpenAIChatCompletionS
2525
)
2626

2727
// private val modelId = NonOpenAIModelId.deepseek_r1
28-
private val modelId = NonOpenAIModelId.llama_v3p1_405b_instruct
28+
private val modelId = NonOpenAIModelId.llama4_maverick_instruct_basic
2929

3030
override protected def run: Future[_] =
3131
service

openai-examples/src/main/scala/io/cequence/openaiscala/examples/fireworksai/FireworksAICreateChatCompletionStreamed.scala

Lines changed: 2 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -21,7 +21,7 @@ object FireworksAICreateChatCompletionStreamed
2121
UserMessage("What is the weather like in Norway?")
2222
)
2323

24-
private val modelId = NonOpenAIModelId.llama_v3p3_70b_instruct
24+
private val modelId = NonOpenAIModelId.llama4_scout_instruct_basic
2525

2626
override protected def run: Future[_] =
2727
service
@@ -37,8 +37,7 @@ object FireworksAICreateChatCompletionStreamed
3737
)
3838
.runWith(
3939
Sink.foreach { completion =>
40-
val content = completion.choices.headOption.flatMap(_.delta.content)
41-
print(content.getOrElse(""))
40+
print(completion.contentHead.getOrElse(""))
4241
}
4342
)
4443
}

openai-examples/src/main/scala/io/cequence/openaiscala/examples/groq/GroqCreateChatCompletion.scala

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -19,7 +19,7 @@ object GroqCreateChatCompletion extends ExampleBase[OpenAIChatCompletionService]
1919
UserMessage("What is the weather like in Norway?")
2020
)
2121

22-
private val modelId = NonOpenAIModelId.llama_3_3_70b_versatile
22+
private val modelId = NonOpenAIModelId.llama_4_scout_17b_16e_instruct
2323

2424
override protected def run: Future[_] =
2525
service
@@ -28,7 +28,7 @@ object GroqCreateChatCompletion extends ExampleBase[OpenAIChatCompletionService]
2828
settings = CreateChatCompletionSettings(
2929
model = modelId,
3030
temperature = Some(0.1),
31-
max_tokens = Some(1024)
31+
max_tokens = Some(2000)
3232
)
3333
)
3434
.map(printMessageContent)

0 commit comments

Comments
 (0)