This reverts commit 6ff2fcc2a7.
This commit was merged in pull request #51.
This commit is contained in:
committed by
GitHub
parent
6ff2fcc2a7
commit
830d1933b1
@@ -5,14 +5,7 @@
|
|||||||
# https://makersuite.google.com/app/apikey
|
# https://makersuite.google.com/app/apikey
|
||||||
GOOGLE_API_KEY=
|
GOOGLE_API_KEY=
|
||||||
|
|
||||||
# https://platform.openai.com/settings/organization/api-keys
|
|
||||||
OPENAI_API_KEY=
|
|
||||||
# https://admin.mistral.ai/organization/api-keys
|
|
||||||
MISTRAL_API_KEY=
|
|
||||||
# https://platform.deepseek.com/api_keys
|
|
||||||
DEEPSEEK_API_KEY=
|
|
||||||
# https://console.x.ai/team/%TEAM_ID%/api-keys
|
|
||||||
XAI_API_KEY=
|
|
||||||
###############################################################################
|
###############################################################################
|
||||||
# Configurazioni per gli agenti di mercato
|
# Configurazioni per gli agenti di mercato
|
||||||
###############################################################################
|
###############################################################################
|
||||||
|
|||||||
12
configs.yaml
12
configs.yaml
@@ -19,18 +19,6 @@ models:
|
|||||||
label: Gemini
|
label: Gemini
|
||||||
# - name: gemini-2.0-pro # TODO Non funziona, ha un nome diverso
|
# - name: gemini-2.0-pro # TODO Non funziona, ha un nome diverso
|
||||||
# label: Gemini Pro
|
# label: Gemini Pro
|
||||||
gpt:
|
|
||||||
- name: gpt-4o
|
|
||||||
label: OpenAIChat
|
|
||||||
deepseek:
|
|
||||||
- name: deepseek-chat
|
|
||||||
label: DeepSeek
|
|
||||||
xai:
|
|
||||||
- name: grok-3
|
|
||||||
label: xAI
|
|
||||||
mistral:
|
|
||||||
- name: mistral-large-latest
|
|
||||||
label: Mistral
|
|
||||||
ollama:
|
ollama:
|
||||||
- name: gpt-oss:latest
|
- name: gpt-oss:latest
|
||||||
label: Ollama GPT
|
label: Ollama GPT
|
||||||
|
|||||||
@@ -24,10 +24,6 @@ dependencies = [
|
|||||||
# Modelli supportati e installati (aggiungere qui sotto quelli che si vogliono usare)
|
# Modelli supportati e installati (aggiungere qui sotto quelli che si vogliono usare)
|
||||||
"google-genai",
|
"google-genai",
|
||||||
"ollama",
|
"ollama",
|
||||||
"openai",
|
|
||||||
"mistralai",
|
|
||||||
"deepseek",
|
|
||||||
"xai",
|
|
||||||
|
|
||||||
# API di exchange di criptovalute
|
# API di exchange di criptovalute
|
||||||
"coinbase-advanced-py",
|
"coinbase-advanced-py",
|
||||||
|
|||||||
@@ -10,10 +10,6 @@ from agno.tools import Toolkit
|
|||||||
from agno.models.base import Model
|
from agno.models.base import Model
|
||||||
from agno.models.google import Gemini
|
from agno.models.google import Gemini
|
||||||
from agno.models.ollama import Ollama
|
from agno.models.ollama import Ollama
|
||||||
from agno.models.openai import OpenAIChat
|
|
||||||
from agno.models.mistral import MistralChat
|
|
||||||
from agno.models.deepseek import DeepSeek
|
|
||||||
from agno.models.xai import xAI
|
|
||||||
|
|
||||||
log = logging.getLogger(__name__)
|
log = logging.getLogger(__name__)
|
||||||
|
|
||||||
@@ -23,6 +19,7 @@ class AppModel(BaseModel):
|
|||||||
name: str = "gemini-2.0-flash"
|
name: str = "gemini-2.0-flash"
|
||||||
label: str = "Gemini"
|
label: str = "Gemini"
|
||||||
model: type[Model] | None = None
|
model: type[Model] | None = None
|
||||||
|
|
||||||
def get_model(self, instructions: str) -> Model:
|
def get_model(self, instructions: str) -> Model:
|
||||||
"""
|
"""
|
||||||
Restituisce un'istanza del modello specificato.
|
Restituisce un'istanza del modello specificato.
|
||||||
@@ -57,84 +54,22 @@ class AppModel(BaseModel):
|
|||||||
output_schema=output_schema
|
output_schema=output_schema
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
class APIConfig(BaseModel):
|
class APIConfig(BaseModel):
|
||||||
retry_attempts: int = 3
|
retry_attempts: int = 3
|
||||||
retry_delay_seconds: int = 2
|
retry_delay_seconds: int = 2
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
class Strategy(BaseModel):
|
class Strategy(BaseModel):
|
||||||
name: str = "Conservative"
|
name: str = "Conservative"
|
||||||
label: str = "Conservative"
|
label: str = "Conservative"
|
||||||
description: str = "Focus on low-risk investments with steady returns."
|
description: str = "Focus on low-risk investments with steady returns."
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
class ModelsConfig(BaseModel):
|
class ModelsConfig(BaseModel):
|
||||||
gemini: list[AppModel] = [AppModel()]
|
gemini: list[AppModel] = [AppModel()]
|
||||||
gpt: list[AppModel] = [AppModel(name="gpt-4o", label="OpenAIChat")]
|
|
||||||
mistral: list[AppModel] = [AppModel(name="mistral-large-latest", label="Mistral")]
|
|
||||||
deepseek: list[AppModel] = [AppModel(name="deepseek-chat", label="DeepSeek")]
|
|
||||||
xai: list[AppModel] = [AppModel(name="grok-3", label="xAI")]
|
|
||||||
ollama: list[AppModel] = []
|
ollama: list[AppModel] = []
|
||||||
|
|
||||||
@property
|
@property
|
||||||
def all_models(self) -> list[AppModel]:
|
def all_models(self) -> list[AppModel]:
|
||||||
return self.gemini + self.ollama + self.gpt + self.mistral + self.deepseek + self.xai
|
return self.gemini + self.ollama
|
||||||
|
|
||||||
def validate_models(self) -> None:
|
|
||||||
"""
|
|
||||||
Validate the configured models for each provider.
|
|
||||||
"""
|
|
||||||
self.__validate_online_models(self.gemini, clazz=Gemini, key="GOOGLE_API_KEY")
|
|
||||||
self.__validate_online_models(self.gpt, clazz=OpenAIChat, key="OPENAI_API_KEY")
|
|
||||||
self.__validate_online_models(self.mistral, clazz=MistralChat, key="MISTRAL_API_KEY")
|
|
||||||
self.__validate_online_models(self.deepseek, clazz=DeepSeek, key="DEEPSEEK_API_KEY")
|
|
||||||
self.__validate_online_models(self.xai, clazz=xAI, key="XAI_API_KEY")
|
|
||||||
|
|
||||||
self.__validate_ollama_models()
|
|
||||||
|
|
||||||
def __validate_online_models(self, models: list[AppModel], clazz: type[Model], key: str | None = None) -> None:
|
|
||||||
"""
|
|
||||||
Validate models for online providers like Gemini.
|
|
||||||
Args:
|
|
||||||
models: list of AppModel instances to validate
|
|
||||||
clazz: class of the model (e.g. Gemini)
|
|
||||||
key: API key required for the provider (optional)
|
|
||||||
"""
|
|
||||||
if key and os.getenv(key) is None:
|
|
||||||
log.warning(f"No {key} set in environment variables for provider.")
|
|
||||||
models.clear()
|
|
||||||
return
|
|
||||||
|
|
||||||
for model in models:
|
|
||||||
model.model = clazz
|
|
||||||
|
|
||||||
def __validate_ollama_models(self) -> None:
|
|
||||||
"""
|
|
||||||
Validate models for the Ollama provider.
|
|
||||||
"""
|
|
||||||
try:
|
|
||||||
models_list = ollama.list()
|
|
||||||
availables = {model['model'] for model in models_list['models']}
|
|
||||||
not_availables: list[str] = []
|
|
||||||
|
|
||||||
for model in self.ollama:
|
|
||||||
if model.name in availables:
|
|
||||||
model.model = Ollama
|
|
||||||
else:
|
|
||||||
not_availables.append(model.name)
|
|
||||||
if not_availables:
|
|
||||||
log.warning(f"Ollama models not available: {not_availables}")
|
|
||||||
|
|
||||||
self.ollama = [model for model in self.ollama if model.model]
|
|
||||||
|
|
||||||
except Exception as e:
|
|
||||||
log.warning(f"Ollama is not running or not reachable: {e}")
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
class AgentsConfigs(BaseModel):
|
class AgentsConfigs(BaseModel):
|
||||||
strategy: str = "Conservative"
|
strategy: str = "Conservative"
|
||||||
@@ -183,7 +118,7 @@ class AppConfig(BaseModel):
|
|||||||
|
|
||||||
super().__init__(*args, **kwargs)
|
super().__init__(*args, **kwargs)
|
||||||
self.set_logging_level()
|
self.set_logging_level()
|
||||||
self.models.validate_models()
|
self.validate_models()
|
||||||
self._initialized = True
|
self._initialized = True
|
||||||
|
|
||||||
def get_model_by_name(self, name: str) -> AppModel:
|
def get_model_by_name(self, name: str) -> AppModel:
|
||||||
@@ -251,3 +186,53 @@ class AppConfig(BaseModel):
|
|||||||
logger = logging.getLogger(logger_name)
|
logger = logging.getLogger(logger_name)
|
||||||
logger.handlers.clear()
|
logger.handlers.clear()
|
||||||
logger.propagate = True
|
logger.propagate = True
|
||||||
|
|
||||||
|
def validate_models(self) -> None:
|
||||||
|
"""
|
||||||
|
Validate the configured models for each provider.
|
||||||
|
"""
|
||||||
|
self.__validate_online_models("gemini", clazz=Gemini, key="GOOGLE_API_KEY")
|
||||||
|
self.__validate_ollama_models()
|
||||||
|
|
||||||
|
def __validate_online_models(self, provider: str, clazz: type[Model], key: str | None = None) -> None:
|
||||||
|
"""
|
||||||
|
Validate models for online providers like Gemini.
|
||||||
|
Args:
|
||||||
|
provider: name of the provider (e.g. "gemini")
|
||||||
|
clazz: class of the model (e.g. Gemini)
|
||||||
|
key: API key required for the provider (optional)
|
||||||
|
"""
|
||||||
|
if getattr(self.models, provider) is None:
|
||||||
|
log.warning(f"No models configured for provider '{provider}'.")
|
||||||
|
|
||||||
|
models: list[AppModel] = getattr(self.models, provider)
|
||||||
|
if key and os.getenv(key) is None:
|
||||||
|
log.warning(f"No {key} set in environment variables for {provider}.")
|
||||||
|
models.clear()
|
||||||
|
return
|
||||||
|
|
||||||
|
for model in models:
|
||||||
|
model.model = clazz
|
||||||
|
|
||||||
|
def __validate_ollama_models(self) -> None:
|
||||||
|
"""
|
||||||
|
Validate models for the Ollama provider.
|
||||||
|
"""
|
||||||
|
try:
|
||||||
|
models_list = ollama.list()
|
||||||
|
availables = {model['model'] for model in models_list['models']}
|
||||||
|
not_availables: list[str] = []
|
||||||
|
|
||||||
|
for model in self.models.ollama:
|
||||||
|
if model.name in availables:
|
||||||
|
model.model = Ollama
|
||||||
|
else:
|
||||||
|
not_availables.append(model.name)
|
||||||
|
if not_availables:
|
||||||
|
log.warning(f"Ollama models not available: {not_availables}")
|
||||||
|
|
||||||
|
self.models.ollama = [model for model in self.models.ollama if model.model]
|
||||||
|
|
||||||
|
except Exception as e:
|
||||||
|
log.warning(f"Ollama is not running or not reachable: {e}")
|
||||||
|
|
||||||
|
|||||||
@@ -83,15 +83,13 @@ class ChatManager:
|
|||||||
label="Modello da usare"
|
label="Modello da usare"
|
||||||
)
|
)
|
||||||
provider.change(fn=self.inputs.choose_team_leader, inputs=provider, outputs=None)
|
provider.change(fn=self.inputs.choose_team_leader, inputs=provider, outputs=None)
|
||||||
provider.value = self.inputs.team_leader_model.label
|
|
||||||
|
|
||||||
strategy = gr.Dropdown(
|
style = gr.Dropdown(
|
||||||
choices=self.inputs.list_strategies_names(),
|
choices=self.inputs.list_strategies_names(),
|
||||||
type="index",
|
type="index",
|
||||||
label="Stile di investimento"
|
label="Stile di investimento"
|
||||||
)
|
)
|
||||||
strategy.change(fn=self.inputs.choose_strategy, inputs=strategy, outputs=None)
|
style.change(fn=self.inputs.choose_strategy, inputs=style, outputs=None)
|
||||||
strategy.value = self.inputs.strategy.label
|
|
||||||
|
|
||||||
chat = gr.ChatInterface(
|
chat = gr.ChatInterface(
|
||||||
fn=self.gradio_respond
|
fn=self.gradio_respond
|
||||||
|
|||||||
Reference in New Issue
Block a user