Add new model configurations and enforce max response length in prompts
This commit is contained in:
12
configs.yaml
12
configs.yaml
@@ -28,6 +28,10 @@ models:
|
||||
label: Qwen 3 (4B)
|
||||
- name: qwen3:1.7b
|
||||
label: Qwen 3 (1.7B)
|
||||
- name: qwen3:32b
|
||||
label: Qwen 3 (32B)
|
||||
- name: phi4-mini:3.8b
|
||||
label: Phi 4 mini (3.8b)
|
||||
|
||||
api:
|
||||
retry_attempts: 3
|
||||
@@ -40,7 +44,7 @@ api:
|
||||
|
||||
agents:
|
||||
strategy: Conservative
|
||||
team_model: qwen3:8b
|
||||
team_leader_model: qwen3:8b
|
||||
query_analyzer_model: qwen3:8b # query check
|
||||
report_generation_model: qwen3:8b # ex predictor
|
||||
team_model: qwen3:32b # the agents
|
||||
team_leader_model: qwen3:32b # the team leader
|
||||
query_analyzer_model: qwen3:32b # query check
|
||||
report_generation_model: qwen3:32b # ex predictor
|
||||
|
||||
Reference in New Issue
Block a user