Skip to content

Commit efec553

Browse files
authored
Update navigate_mxgenai.md
small changes to text
1 parent 0d3db87 commit efec553

File tree

1 file changed

+3
-3
lines changed
  • content/en/docs/appstore/use-content/platform-supported-content/modules/genai/mendix-cloud-genai

1 file changed

+3
-3
lines changed

content/en/docs/appstore/use-content/platform-supported-content/modules/genai/mendix-cloud-genai/navigate_mxgenai.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -93,12 +93,12 @@ The Token consumption monitor shows detailed graphs of the token consumption use
9393

9494
### Why do we measure token consumption?
9595

96-
In order for a large language model to understand text input, the text is first ‘tokenized’ - broken down into smaller pieces where each piece represents a token with its unique ID. A good rule of thumb is that 100 tokens are around 75 English words, however there are always differences depending on the model or the language used. After tokenization, each token will be assigned an embeddings vector. The tokens required to feed the input prompt to the model are called ‘input tokens’, the tokens required to transform the model output into for example text or images are called ‘output tokens’. As a result, only input token consumption is being measured for Embedding resources, since only the generated embedding vectors are returned and no tokenization takes place. Text generation resources contain both input and output tokens (text sent to the model and generated by the model).
96+
In order for a large language model to understand text input, the text is first ‘tokenized’ - broken down into smaller pieces where each piece represents a token with its unique ID. A good rule of thumb is that 100 tokens are around 75 English words, however there are always differences depending on the model or the language used. After tokenization, each token will be assigned an embeddings vector. The tokens required to feed the input prompt to the model are called ‘input tokens’, the tokens required to transform the model output into for example text or images are called ‘output tokens’. Tokens are what you pay for when consuming large language model services. For Embeddings resources, only input token consumption is being measured, since only the generated embedding vectors are returned and no tokenization takes place when generating the output. Text generation resources contain both input and output tokens (text sent to the model and generated by the model).
9797

9898
### Knowledgebases and Embeddings Resources
9999

100-
A Knowledgebase only stores embeddings, and uses a Embedding resources to embed files. If there is a knowledgebase connected to the embedding resource, uploading a document in the knowledgebase will consume tokens in the embeddings resource.
100+
A Knowledgebase only stores embedding vectors, and uses an Embeddings resource to calculate them. If there is a knowledgebase connected to the Embeddings resource, uploading a document in the knowledgebase will consume tokens in the embeddings resource.
101101

102102
### Exporting Token Consumption Data
103103

104-
It is possible to export consumption data to CSV. The export contains basic information about input tokens, output tokens and date. Days with no consumption are not exported.
104+
It is possible to export consumption data to CSV. The export contains basic information about input tokens, output tokens and dates. Days with no consumption are not exported.

0 commit comments

Comments
 (0)