OpenAI compatible web server: chat completion **streaming**: **missing spaces** #1208
-
The web server is started with: llama_cpp_config.json:
client = openai.OpenAI( messages = [
response_chat = client.chat.completions.create( response_chat:
response_stream = client.chat.completions.create( response_contents = [] ChatCompletionChunk(id='chatcmpl-583c9ff9-6d34', choices=[Choice(delta=ChoiceDelta(content=None, ...) full_response = ''.join([m for m in response_contents if m is not None]) full_response: ** Issue **
Thanks a lot for any input, comment, suggestion... |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment 1 reply
-
@didierguillevic there was a bug with the LlamaHFTokenizer but it should be fixed in v0.2.50 |
Beta Was this translation helpful? Give feedback.
@didierguillevic there was a bug with the LlamaHFTokenizer but it should be fixed in v0.2.50