Skip to content

Do you support streaming generating outputs? #245

Answered by WoosukKwon
ltz0120 asked this question in Q&A
Discussion options

You must be logged in to vote

Yes, our FastAPI and OpenAI servers support streaming outputs. Just set up the server with

python -m vllm.entrypoints.api_server

or

python -m vllm.entrypoints.openai.api_server

and then add "stream": True in client request (by default it's false).
See

"stream": stream,

Replies: 5 comments

Comment options

You must be logged in to vote
0 replies
Answer selected by zhuohan123
Comment options

You must be logged in to vote
0 replies
Comment options

You must be logged in to vote
0 replies
Comment options

You must be logged in to vote
0 replies
Comment options

You must be logged in to vote
0 replies
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
question Further information is requested
6 participants
Converted from issue

This discussion was converted from issue #230 on June 25, 2023 17:47.