You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: docs/providers/ai21.md
+7-7
Original file line number
Diff line number
Diff line change
@@ -46,14 +46,14 @@ The following model aliases are provided for this provider.
46
46
47
47
The following parameters can be passed through `options`.
48
48
49
-
-`logprobs`: _Details not available, please refer to the LLM provider documentation._
50
-
-`max_tokens`: _Details not available, please refer to the LLM provider documentation._
51
-
-`n`: _Details not available, please refer to the LLM provider documentation._
52
-
-`stop`: _Details not available, please refer to the LLM provider documentation._
53
-
-`stream`: _Details not available, please refer to the LLM provider documentation._
54
-
-`temperature`: _Details not available, please refer to the LLM provider documentation._
49
+
-`logprobs`: Includes the log probabilities of the most likely tokens, providing insights into the model's token selection process.
50
+
-`max_tokens`: The maximum number of tokens that can be generated in the chat completion. The total length of input tokens and generated tokens is limited by the model's context length.
51
+
-`n`: Specifies the number of responses to generate for each input message. Note that costs are based on the number of generated tokens across all choices. Keeping n as 1 minimizes costs.
52
+
-`stop`: Up to 4 sequences where the API will stop generating further tokens.
53
+
-`stream`: If set, partial message deltas will be sent, similar to ChatGPT. Tokens will be sent as data-only server-sent events as they become available, with the stream terminated by a data: [DONE] message.
54
+
-`temperature`: Controls the randomness of the AI's responses. A higher temperature results in more random outputs, while a lower temperature makes the output more focused and deterministic. Generally, it is recommended to alter this or top_p, but not both.
55
55
-`top_logprobs`: _Details not available, please refer to the LLM provider documentation._
56
-
-`top_p`: _Details not available, please refer to the LLM provider documentation._
56
+
-`top_p`: Controls the cumulative probability of token selections for nucleus sampling. It limits the tokens to the smallest set whose cumulative probability exceeds the threshold. It is recommended to alter this or temperature, but not both.
Copy file name to clipboardExpand all lines: docs/providers/ailayer.md
+2-2
Original file line number
Diff line number
Diff line change
@@ -42,8 +42,8 @@ The following model aliases are provided for this provider.
42
42
43
43
The following parameters can be passed through `options`.
44
44
45
-
-`max_tokens`: _Details not available, please refer to the LLM provider documentation._
46
-
-`temperature`: _Details not available, please refer to the LLM provider documentation._
45
+
-`max_tokens`: The maximum number of tokens that can be generated in the chat completion. The total length of input tokens and generated tokens is limited by the model's context length.
46
+
-`temperature`: Controls the randomness of the AI's responses. A higher temperature results in more random outputs, while a lower temperature makes the output more focused and deterministic. Generally, it is recommended to alter this or top_p, but not both.
Copy file name to clipboardExpand all lines: docs/providers/aimlapi.md
+5-5
Original file line number
Diff line number
Diff line change
@@ -48,11 +48,11 @@ The following model aliases are provided for this provider.
48
48
49
49
The following parameters can be passed through `options`.
50
50
51
-
-`frequency_penalty`: _Details not available, please refer to the LLM provider documentation._
52
-
-`max_tokens`: _Details not available, please refer to the LLM provider documentation._
53
-
-`stream`: _Details not available, please refer to the LLM provider documentation._
54
-
-`temperature`: _Details not available, please refer to the LLM provider documentation._
55
-
-`top_p`: _Details not available, please refer to the LLM provider documentation._
51
+
-`frequency_penalty`: Penalizes new tokens based on their existing frequency in the text so far, reducing the likelihood of repeating the same line. Positive values reduce the frequency of tokens appearing in the generated text.
52
+
-`max_tokens`: The maximum number of tokens that can be generated in the chat completion. The total length of input tokens and generated tokens is limited by the model's context length.
53
+
-`stream`: If set, partial message deltas will be sent, similar to ChatGPT. Tokens will be sent as data-only server-sent events as they become available, with the stream terminated by a data: [DONE] message.
54
+
-`temperature`: Controls the randomness of the AI's responses. A higher temperature results in more random outputs, while a lower temperature makes the output more focused and deterministic. Generally, it is recommended to alter this or top_p, but not both.
55
+
-`top_p`: Controls the cumulative probability of token selections for nucleus sampling. It limits the tokens to the smallest set whose cumulative probability exceeds the threshold. It is recommended to alter this or temperature, but not both.
Copy file name to clipboardExpand all lines: docs/providers/anthropic.md
+10-10
Original file line number
Diff line number
Diff line change
@@ -42,16 +42,16 @@ The following model aliases are provided for this provider.
42
42
43
43
The following parameters can be passed through `options`.
44
44
45
-
-`max_tokens`: _Details not available, please refer to the LLM provider documentation._
46
-
-`metadata`: _Details not available, please refer to the LLM provider documentation._
47
-
-`stop_sequences`: _Details not available, please refer to the LLM provider documentation._
48
-
-`stream`: _Details not available, please refer to the LLM provider documentation._
49
-
-`system`: _Details not available, please refer to the LLM provider documentation._
50
-
-`temperature`: _Details not available, please refer to the LLM provider documentation._
51
-
-`tool_choice`: _Details not available, please refer to the LLM provider documentation._
52
-
-`tools`: _Details not available, please refer to the LLM provider documentation._
53
-
-`top_k`: _Details not available, please refer to the LLM provider documentation._
54
-
-`top_p`: _Details not available, please refer to the LLM provider documentation._
45
+
-`max_tokens`: The maximum number of tokens that can be generated in the chat completion. The total length of input tokens and generated tokens is limited by the model's context length.
46
+
-`metadata`: Additional information about the input or environment that might influence the AI's response.
47
+
-`stop_sequences`: Sequences that indicate to the model when to stop generating further tokens.
48
+
-`stream`: If set, partial message deltas will be sent, similar to ChatGPT. Tokens will be sent as data-only server-sent events as they become available, with the stream terminated by a data: [DONE] message.
49
+
-`system`: Defines the role and instructions for the system component of the AI interaction, guiding the overall behavior.
50
+
-`temperature`: Controls the randomness of the AI's responses. A higher temperature results in more random outputs, while a lower temperature makes the output more focused and deterministic. Generally, it is recommended to alter this or top_p, but not both.
51
+
-`tool_choice`: Specifies which external tools the AI can use to assist in generating its response.
52
+
-`tools`: A list of external tools available for the AI to use in generating responses.
53
+
-`top_k`: The number of highest probability vocabulary tokens to keep for top-k sampling.
54
+
-`top_p`: Controls the cumulative probability of token selections for nucleus sampling. It limits the tokens to the smallest set whose cumulative probability exceeds the threshold. It is recommended to alter this or temperature, but not both.
Copy file name to clipboardExpand all lines: docs/providers/anyscale.md
+4-4
Original file line number
Diff line number
Diff line change
@@ -48,10 +48,10 @@ The following model aliases are provided for this provider.
48
48
49
49
The following parameters can be passed through `options`.
50
50
51
-
-`frequency_penalty`: _Details not available, please refer to the LLM provider documentation._
52
-
-`max_tokens`: _Details not available, please refer to the LLM provider documentation._
53
-
-`temperature`: _Details not available, please refer to the LLM provider documentation._
54
-
-`top_p`: _Details not available, please refer to the LLM provider documentation._
51
+
-`frequency_penalty`: Penalizes new tokens based on their existing frequency in the text so far, reducing the likelihood of repeating the same line. Positive values reduce the frequency of tokens appearing in the generated text.
52
+
-`max_tokens`: The maximum number of tokens that can be generated in the chat completion. The total length of input tokens and generated tokens is limited by the model's context length.
53
+
-`temperature`: Controls the randomness of the AI's responses. A higher temperature results in more random outputs, while a lower temperature makes the output more focused and deterministic. Generally, it is recommended to alter this or top_p, but not both.
54
+
-`top_p`: Controls the cumulative probability of token selections for nucleus sampling. It limits the tokens to the smallest set whose cumulative probability exceeds the threshold. It is recommended to alter this or temperature, but not both.
Copy file name to clipboardExpand all lines: docs/providers/cloudflareai.md
+2-2
Original file line number
Diff line number
Diff line change
@@ -48,8 +48,8 @@ The following model aliases are provided for this provider.
48
48
49
49
The following parameters can be passed through `options`.
50
50
51
-
-`max_tokens`: _Details not available, please refer to the LLM provider documentation._
52
-
-`temperature`: _Details not available, please refer to the LLM provider documentation._
51
+
-`max_tokens`: The maximum number of tokens that can be generated in the chat completion. The total length of input tokens and generated tokens is limited by the model's context length.
52
+
-`temperature`: Controls the randomness of the AI's responses. A higher temperature results in more random outputs, while a lower temperature makes the output more focused and deterministic. Generally, it is recommended to alter this or top_p, but not both.
Copy file name to clipboardExpand all lines: docs/providers/cohere.md
+8-8
Original file line number
Diff line number
Diff line change
@@ -53,20 +53,20 @@ The following parameters can be passed through `options`.
53
53
-`conversation_id`: _Details not available, please refer to the LLM provider documentation._
54
54
-`documents`: _Details not available, please refer to the LLM provider documentation._
55
55
-`force_single_step`: _Details not available, please refer to the LLM provider documentation._
56
-
-`frequency_penalty`: _Details not available, please refer to the LLM provider documentation._
56
+
-`frequency_penalty`: Penalizes new tokens based on their existing frequency in the text so far, reducing the likelihood of repeating the same line. Positive values reduce the frequency of tokens appearing in the generated text.
57
57
-`k`: _Details not available, please refer to the LLM provider documentation._
58
58
-`max_input_tokens`: _Details not available, please refer to the LLM provider documentation._
59
-
-`max_tokens`: _Details not available, please refer to the LLM provider documentation._
59
+
-`max_tokens`: The maximum number of tokens that can be generated in the chat completion. The total length of input tokens and generated tokens is limited by the model's context length.
60
60
-`p`: _Details not available, please refer to the LLM provider documentation._
61
61
-`preamble`: _Details not available, please refer to the LLM provider documentation._
62
-
-`presence_penalty`: _Details not available, please refer to the LLM provider documentation._
62
+
-`presence_penalty`: Penalizes new tokens based on whether they appear in the text so far, encouraging the model to talk about new topics. Positive values increase the likelihood of new tokens appearing in the generated text.
63
63
-`prompt_truncation`: _Details not available, please refer to the LLM provider documentation._
64
-
-`seed`: _Details not available, please refer to the LLM provider documentation._
65
-
-`stop_sequences`: _Details not available, please refer to the LLM provider documentation._
66
-
-`stream`: _Details not available, please refer to the LLM provider documentation._
67
-
-`temperature`: _Details not available, please refer to the LLM provider documentation._
64
+
-`seed`: A random seed for reproducibility. If specified, the system will attempt to sample deterministically, ensuring repeated requests with the same seed and parameters return the same result. Determinism is not guaranteed.
65
+
-`stop_sequences`: Sequences that indicate to the model when to stop generating further tokens.
66
+
-`stream`: If set, partial message deltas will be sent, similar to ChatGPT. Tokens will be sent as data-only server-sent events as they become available, with the stream terminated by a data: [DONE] message.
67
+
-`temperature`: Controls the randomness of the AI's responses. A higher temperature results in more random outputs, while a lower temperature makes the output more focused and deterministic. Generally, it is recommended to alter this or top_p, but not both.
68
68
-`tool_results`: _Details not available, please refer to the LLM provider documentation._
69
-
-`tools`: _Details not available, please refer to the LLM provider documentation._
69
+
-`tools`: A list of external tools available for the AI to use in generating responses.
Copy file name to clipboardExpand all lines: docs/providers/corcel.md
+4-4
Original file line number
Diff line number
Diff line change
@@ -42,10 +42,10 @@ The following model aliases are provided for this provider.
42
42
43
43
The following parameters can be passed through `options`.
44
44
45
-
-`max_tokens`: _Details not available, please refer to the LLM provider documentation._
46
-
-`stream`: _Details not available, please refer to the LLM provider documentation._
47
-
-`temperature`: _Details not available, please refer to the LLM provider documentation._
48
-
-`top_p`: _Details not available, please refer to the LLM provider documentation._
45
+
-`max_tokens`: The maximum number of tokens that can be generated in the chat completion. The total length of input tokens and generated tokens is limited by the model's context length.
46
+
-`stream`: If set, partial message deltas will be sent, similar to ChatGPT. Tokens will be sent as data-only server-sent events as they become available, with the stream terminated by a data: [DONE] message.
47
+
-`temperature`: Controls the randomness of the AI's responses. A higher temperature results in more random outputs, while a lower temperature makes the output more focused and deterministic. Generally, it is recommended to alter this or top_p, but not both.
48
+
-`top_p`: Controls the cumulative probability of token selections for nucleus sampling. It limits the tokens to the smallest set whose cumulative probability exceeds the threshold. It is recommended to alter this or temperature, but not both.
@@ -48,18 +48,18 @@ The following model aliases are provided for this provider.
48
48
49
49
The following parameters can be passed through `options`.
50
50
51
-
-`echo`: _Details not available, please refer to the LLM provider documentation._
52
-
-`frequency_penalty`: _Details not available, please refer to the LLM provider documentation._
53
-
-`max_tokens`: _Details not available, please refer to the LLM provider documentation._
54
-
-`n`: _Details not available, please refer to the LLM provider documentation._
55
-
-`presence_penalty`: _Details not available, please refer to the LLM provider documentation._
56
-
-`response_format`: _Details not available, please refer to the LLM provider documentation._
57
-
-`stop`: _Details not available, please refer to the LLM provider documentation._
58
-
-`stream`: _Details not available, please refer to the LLM provider documentation._
59
-
-`temperature`: _Details not available, please refer to the LLM provider documentation._
60
-
-`tool_choice`: _Details not available, please refer to the LLM provider documentation._
61
-
-`tools`: _Details not available, please refer to the LLM provider documentation._
62
-
-`top_p`: _Details not available, please refer to the LLM provider documentation._
51
+
-`echo`: If set to true, the input prompt is echoed back in the output.
52
+
-`frequency_penalty`: Penalizes new tokens based on their existing frequency in the text so far, reducing the likelihood of repeating the same line. Positive values reduce the frequency of tokens appearing in the generated text.
53
+
-`max_tokens`: The maximum number of tokens that can be generated in the chat completion. The total length of input tokens and generated tokens is limited by the model's context length.
54
+
-`n`: Specifies the number of responses to generate for each input message. Note that costs are based on the number of generated tokens across all choices. Keeping n as 1 minimizes costs.
55
+
-`presence_penalty`: Penalizes new tokens based on whether they appear in the text so far, encouraging the model to talk about new topics. Positive values increase the likelihood of new tokens appearing in the generated text.
56
+
-`response_format`: Defines the format of the AI's response. Setting this to { "type": "json_object" } enables JSON mode, ensuring the message generated by the model is valid JSON.
57
+
-`stop`: Up to 4 sequences where the API will stop generating further tokens.
58
+
-`stream`: If set, partial message deltas will be sent, similar to ChatGPT. Tokens will be sent as data-only server-sent events as they become available, with the stream terminated by a data: [DONE] message.
59
+
-`temperature`: Controls the randomness of the AI's responses. A higher temperature results in more random outputs, while a lower temperature makes the output more focused and deterministic. Generally, it is recommended to alter this or top_p, but not both.
60
+
-`tool_choice`: Specifies which external tools the AI can use to assist in generating its response.
61
+
-`tools`: A list of external tools available for the AI to use in generating responses.
62
+
-`top_p`: Controls the cumulative probability of token selections for nucleus sampling. It limits the tokens to the smallest set whose cumulative probability exceeds the threshold. It is recommended to alter this or temperature, but not both.
63
63
64
64
65
65
### Features
@@ -82,12 +82,3 @@ To get an API key, first create a DeepInfra account, then visit the link below.
0 commit comments