2020
2121import prompt
2222import utils
23- from config import LLM_ENDPOINT , VLM_ENDPOINT , TTS_ACCESS_TOKEN , TTS_APP_ID
23+ from config import VISUAL_SUMMARY_ENDPOINT , QUESTION_ANSWER_ENDPOINT , TTS_ACCESS_TOKEN , TTS_APP_ID
2424
2525from arkitect .core .component .llm import BaseChatLanguageModel
2626from arkitect .core .component .llm .model import (
2929 ArkChatRequest ,
3030 ArkChatResponse ,
3131 ArkMessage ,
32+ ArkThinking ,
3233 ChatCompletionMessageTextPart ,
3334 Response ,
3435)
@@ -74,7 +75,7 @@ async def chat_with_vlm(
7475 parameters : ArkChatParameters ,
7576) -> Tuple [bool , Optional [AsyncIterable [ArkChatCompletionChunk ]]]:
7677 vlm = BaseChatLanguageModel (
77- endpoint_id = VLM_ENDPOINT ,
78+ endpoint_id = VISUAL_SUMMARY_ENDPOINT ,
7879 messages = [ArkMessage (role = "system" , content = prompt .VLM_CHAT_PROMPT )]
7980 + [request .messages [- 1 ]],
8081 parameters = parameters ,
@@ -108,7 +109,7 @@ async def llm_answer(
108109 contexts , context_id , request , prompt .LLM_PROMPT
109110 )
110111 llm = BaseChatLanguageModel (
111- endpoint_id = LLM_ENDPOINT ,
112+ endpoint_id = QUESTION_ANSWER_ENDPOINT ,
112113 messages = request_messages ,
113114 parameters = parameters ,
114115 )
@@ -180,7 +181,7 @@ async def summarize_image(
180181 ArkMessage (role = "system" , content = prompt .VLM_PROMPT )
181182 ] + request .messages
182183 vlm = BaseChatLanguageModel (
183- endpoint_id = VLM_ENDPOINT ,
184+ endpoint_id = VISUAL_SUMMARY_ENDPOINT ,
184185 messages = request_messages ,
185186 parameters = parameters ,
186187 )
@@ -195,7 +196,7 @@ async def default_model_calling(
195196 request : ArkChatRequest ,
196197) -> AsyncIterable [Union [ArkChatCompletionChunk , ArkChatResponse ]]:
197198 # local in-memory storage should be changed to other storage in production
198- context_id : Optional [str ] = get_headers (). get ( "X-Context-Id" , None )
199+ context_id : Optional [str ] = request . metadata [ "context_id" ]
199200 assert context_id is not None
200201 contexts : utils .Storage = utils .CoroutineSafeMap .get_instance_sync ()
201202 if not await contexts .contains (context_id ):
@@ -209,6 +210,7 @@ async def default_model_calling(
209210 and request .messages [- 1 ].content [0 ].text == ""
210211 )
211212 parameters = ArkChatParameters (** request .__dict__ )
213+ parameters .thinking = ArkThinking (type = "disabled" )
212214 if is_image :
213215 _ = asyncio .create_task (
214216 summarize_image (contexts , request , parameters , context_id )
0 commit comments