From 323802e7405758842184a225e8a53e2d351bf989 Mon Sep 17 00:00:00 2001 From: Charlotte Som Date: Wed, 26 Feb 2025 16:13:31 +0000 Subject: [PATCH] use user-supplied model id --- server/inference.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/server/inference.py b/server/inference.py index 1d3a92b..4327aac 100644 --- a/server/inference.py +++ b/server/inference.py @@ -24,7 +24,7 @@ async def connect_to_conversation(ws: WebSocket): continuing = bool(ws.query_params.get("continue")) conversation_id = ws.path_params["conversation"] if conversation_id == "new": - conversation = llm.AsyncConversation(llm.get_async_model()) + conversation = llm.AsyncConversation(llm.get_async_model(ws.query_params.get("model"))) else: try: conversation: llm.AsyncConversation = llm.cli.load_conversation(conversation_id, async_=True)