2020
2121import  prompt 
2222import  utils 
23- from  config  import  LLM_ENDPOINT ,  VLM_ENDPOINT , TTS_ACCESS_TOKEN , TTS_APP_ID 
23+ from  config  import  VISUAL_SUMMARY_ENDPOINT ,  QUESTION_ANSWER_ENDPOINT , TTS_ACCESS_TOKEN , TTS_APP_ID 
2424
2525from  arkitect .core .component .llm  import  BaseChatLanguageModel 
2626from  arkitect .core .component .llm .model  import  (
2929    ArkChatRequest ,
3030    ArkChatResponse ,
3131    ArkMessage ,
32+     ArkThinking ,
3233    ChatCompletionMessageTextPart ,
3334    Response ,
3435)
@@ -74,7 +75,7 @@ async def chat_with_vlm(
7475    parameters : ArkChatParameters ,
7576) ->  Tuple [bool , Optional [AsyncIterable [ArkChatCompletionChunk ]]]:
7677    vlm  =  BaseChatLanguageModel (
77-         endpoint_id = VLM_ENDPOINT ,
78+         endpoint_id = VISUAL_SUMMARY_ENDPOINT ,
7879        messages = [ArkMessage (role = "system" , content = prompt .VLM_CHAT_PROMPT )]
7980        +  [request .messages [- 1 ]],
8081        parameters = parameters ,
@@ -108,7 +109,7 @@ async def llm_answer(
108109        contexts , context_id , request , prompt .LLM_PROMPT 
109110    )
110111    llm  =  BaseChatLanguageModel (
111-         endpoint_id = LLM_ENDPOINT ,
112+         endpoint_id = QUESTION_ANSWER_ENDPOINT ,
112113        messages = request_messages ,
113114        parameters = parameters ,
114115    )
@@ -180,7 +181,7 @@ async def summarize_image(
180181        ArkMessage (role = "system" , content = prompt .VLM_PROMPT )
181182    ] +  request .messages 
182183    vlm  =  BaseChatLanguageModel (
183-         endpoint_id = VLM_ENDPOINT ,
184+         endpoint_id = VISUAL_SUMMARY_ENDPOINT ,
184185        messages = request_messages ,
185186        parameters = parameters ,
186187    )
@@ -195,7 +196,7 @@ async def default_model_calling(
195196    request : ArkChatRequest ,
196197) ->  AsyncIterable [Union [ArkChatCompletionChunk , ArkChatResponse ]]:
197198    # local in-memory storage should be changed to other storage in production 
198-     context_id : Optional [str ] =  get_headers (). get ( "X-Context-Id" ,  None ) 
199+     context_id : Optional [str ] =  request . metadata [ "context_id" ] 
199200    assert  context_id  is  not   None 
200201    contexts : utils .Storage  =  utils .CoroutineSafeMap .get_instance_sync ()
201202    if  not  await  contexts .contains (context_id ):
@@ -209,6 +210,7 @@ async def default_model_calling(
209210        and  request .messages [- 1 ].content [0 ].text  ==  "" 
210211    )
211212    parameters  =  ArkChatParameters (** request .__dict__ )
213+     parameters .thinking  =  ArkThinking (type = "disabled" )
212214    if  is_image :
213215        _  =  asyncio .create_task (
214216            summarize_image (contexts , request , parameters , context_id )
0 commit comments