| # This is an abstract flow, therefore some required fields are not defined (and must be defined by the concrete flow) | |
| _target_: flow_modules.aiflows.ChatFlowModule.ChatAtomicFlow.instantiate_from_default_config | |
| name: ChatAtomicFlow | |
| description: "Flow which uses as tool an LLM though an API" | |
| enable_cache: True | |
| n_api_retries: 6 | |
| wait_time_between_retries: 20 | |
| system_name: system | |
| user_name: user | |
| assistant_name: assistant | |
| backend: | |
| _target_: aiflows.backends.llm_lite.LiteLLMBackend | |
| api_infos: ??? | |
| model_name: | |
| openai: "gpt-3.5-turbo" | |
| n: 1 | |
| max_tokens: 2000 | |
| temperature: 0.3 | |
| top_p: 0.2 | |
| stream: True | |
| system_message_prompt_template: | |
| _target_: aiflows.prompt_template.JinjaPrompt | |
| init_human_message_prompt_template: | |
| _target_: aiflows.prompt_template.JinjaPrompt | |
| template: "{{query}}" | |
| input_variables: | |
| - "query" | |
| human_message_prompt_template: | |
| _target_: aiflows.prompt_template.JinjaPrompt | |
| template: "{{query}}" | |
| input_variables: | |
| - "query" | |
| input_interface_initialized: | |
| - "query" | |
| query_message_prompt_template: | |
| _target_: aiflows.prompt_template.JinjaPrompt | |
| previous_messages: | |
| first_k: null # Note that the first message is the system prompt | |
| last_k: null | |
| output_interface: | |
| - "api_output" | |