【WIP】feat: add local llama-cpp embedding support#1388
【WIP】feat: add local llama-cpp embedding support#1388Mijamind719 wants to merge 3 commits intovolcengine:mainfrom
Conversation
Co-authored-by: GPT-5.4 <noreply@openai.com>
PR Reviewer Guide 🔍Here are some key observations to aid the review process:
|
PR Code Suggestions ✨Explore these optional code suggestions:
|
Co-authored-by: GPT-5.4 <noreply@openai.com>
Legacy issue: investigate true llama-cpp native multi-sequence batch support for local embedding models such as bge-small-zh-v1.5-f16 (current runtime reports n_seq_max=1, so embed_batch uses sequential mode). Co-authored-by: GPT-5.4 <noreply@openai.com>
9c8e7ef to
6e57688
Compare
Co-authored-by: GPT-5.4
Description
Related Issue
Type of Change
Changes Made
Testing
Checklist
Screenshots (if applicable)
Additional Notes