Skip to content

Support JSON mode. #2483

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Closed
MiyazonoKaori opened this issue Jan 18, 2024 · 11 comments
Closed

Support JSON mode. #2483

MiyazonoKaori opened this issue Jan 18, 2024 · 11 comments

Comments

@MiyazonoKaori
Copy link

Any plan to integrate modules such as lm-format-enforcer. Support JSON mode.

@simon-mo
Copy link
Collaborator

Yes! This is our top priorities.

@hadsed
Copy link

hadsed commented Jan 19, 2024

Is there currently any work happening here? I could look into it if not.

@fullstackwebdev
Copy link

context-free grammar
FwDTXcNaYAIafOg

it would be cool if the llama cpp GBNF grammars could be standardize so we can do stuff like the image , between different LLM frameworks.

@simon-mo
Copy link
Collaborator

In our roadmap, anything https://github.com/outlines-dev/outlines supports can be part of vLLM, as long as any other frameworks that support LogitsProcessors API.

@Kaotic3
Copy link

Kaotic3 commented Jan 23, 2024

This looks good, just read through outlines - seems super useful.

When do we get it :D

@findalexli
Copy link

Have someone done benchmark on latency introduced by outlines?

@solesensei
Copy link

Is this already supported as described here #1191?

@ChuckHend
Copy link

I would also love to see this feature make it into vllm!

@wxgeorge
Copy link

wxgeorge commented Aug 16, 2024

Is this already supported as described here #1191?

I suspect the same.

Specifically #3211 implements handling for "request_format": { "type": "json_object" } in the completion request body, which is what I'm familiar with when we say "JSON mode".

@ChuckHend
Copy link

"request_format": { "type": "json_object" } seems to work for me. Also requires a prompt specifically asking for json response.

@DarkLight1337
Copy link
Member

Is this already supported as described here #1191?

I suspect the same.

Specifically #3211 implements handling for "request_format": { "type": "json_object" } in the completion request body, which is what I'm familiar with when we say "JSON mode".

Yep, that should do it.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

10 participants