Performance for very large dataset #663
Unanswered
roldengarm
asked this question in
1. Q&A
Replies: 1 comment 8 replies
-
I'm assuming this is related to #666 A few recommendations:
|
Beta Was this translation helpful? Give feedback.
8 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
We're using Kernel Memory as a service to ingest about 9 million text records. It's set up as a service on an Azure App Service, with Azure Queues, embedding-3-large on Azure OpenAI and Postgres as database.
To ingest, we're using an Azure Function that calls the KM Web Service to ingest a document & waits until it's ready. It's configured to do max 12 in parallel.
The current throughput is about 150-200 text records per minute, so the entire data set will take 30-40 days.
Initially I had it running without any throttling, i.e. the Azure Function would just keep ingesting documents, but then the KM service would fall over, for further details see here. That's when I implemented the parallelization.
In that topic, batching was discussed, but that isn't ready yet.
App Service Plan runs at ~10-20% CPU. The main bottleneck seems to be the embedding generation; when I let it go unthrottled it went down because of quota in Azure OpenAI.
I've just been on the Semantic Kernel Office Hours chat and they recommended to reach out here.
In the interim, is there anything we can do to improve the performance?
Beta Was this translation helpful? Give feedback.
All reactions