@@ -1020,6 +1020,8 @@ async def oai_chat_completions(user_prompt,
10201020 event_data = line .split (': ' , 1 )
10211021 assert event_data [0 ] == 'data' , f'Bad event code received: ```{ event_data } ```'
10221022 chunk_raw = event_data [1 ]
1023+ if chunk_raw == '[DONE]' :
1024+ continue
10231025
10241026 chunk = json .loads (chunk_raw )
10251027 assert len (chunk ['choices' ]) == 1 , f"no choices provided, line ```{ line } ```"
@@ -1322,14 +1324,16 @@ def start_server_background(context):
13221324 if 'LLAMA_SERVER_BIN_PATH' in os .environ :
13231325 context .server_path = os .environ ['LLAMA_SERVER_BIN_PATH' ]
13241326 server_listen_addr = context .server_fqdn
1327+ # server_listen_addr = "127.0.0.1"
13251328 server_args = [
13261329 '--host' , server_listen_addr ,
13271330 '--port' , context .server_port ,
13281331 ]
1332+
13291333 if context .model_file :
13301334 server_args .extend (['--model' , context .model_file ])
1331- if context .model_url :
1332- server_args .extend (['--model-url' , context .model_url ])
1335+ # if context.model_url:
1336+ # server_args.extend(['--model-url', context.model_url])
13331337 if context .model_hf_repo :
13341338 server_args .extend (['--hf-repo' , context .model_hf_repo ])
13351339 if context .model_hf_file :
0 commit comments