mirror of
https://github.com/trycua/computer.git
synced 2026-01-01 02:50:15 -06:00
Restore gemma
This commit is contained in:
@@ -34,7 +34,7 @@ async def run_agent_example():
|
||||
# model=LLM(provider=LLMProvider.OPENAI), # No model name for Operator CUA
|
||||
# model=LLM(provider=LLMProvider.OPENAI, name="gpt-4.5-preview"),
|
||||
# model=LLM(provider=LLMProvider.ANTHROPIC, name="claude-3-7-sonnet-20250219"),
|
||||
model=LLM(provider=LLMProvider.OLLAMA, name="qwen2.5:7b"),
|
||||
model=LLM(provider=LLMProvider.OLLAMA, name="gemma3:4b-it-q4_K_M"),
|
||||
save_trajectory=True,
|
||||
only_n_most_recent_images=3,
|
||||
verbosity=logging.DEBUG,
|
||||
|
||||
@@ -6,7 +6,7 @@ from ..providers.omni.types import LLMProvider
|
||||
DEFAULT_MODELS = {
|
||||
LLMProvider.OPENAI: "gpt-4o",
|
||||
LLMProvider.ANTHROPIC: "claude-3-7-sonnet-20250219",
|
||||
LLMProvider.OLLAMA: "qwen2.5:7b",
|
||||
LLMProvider.OLLAMA: "gemma3:4b-it-q4_K_M",
|
||||
}
|
||||
|
||||
# Map providers to their environment variable names
|
||||
|
||||
@@ -20,7 +20,7 @@ class OllamaClient(BaseOmniClient):
|
||||
|
||||
Args:
|
||||
api_key: Not used
|
||||
model: Ollama model name (e.g. "qwen2.5:7b")
|
||||
model: Ollama model name (e.g. "gemma3:4b-it-q4_K_M")
|
||||
max_retries: Maximum number of retries for API calls
|
||||
retry_delay: Base delay between retries in seconds
|
||||
"""
|
||||
|
||||
@@ -36,7 +36,7 @@ Model = LLM
|
||||
PROVIDER_TO_DEFAULT_MODEL: Dict[LLMProvider, str] = {
|
||||
LLMProvider.ANTHROPIC: "claude-3-7-sonnet-20250219",
|
||||
LLMProvider.OPENAI: "gpt-4o",
|
||||
LLMProvider.OLLAMA: "qwen2.5:7b",
|
||||
LLMProvider.OLLAMA: "gemma3:4b-it-q4_K_M",
|
||||
}
|
||||
|
||||
# Environment variable names for each provider
|
||||
|
||||
Reference in New Issue
Block a user