diff --git a/docs/content/docs/agent-sdk/integrations/hud.mdx b/docs/content/docs/agent-sdk/integrations/hud.mdx index b517121e..cee5f77f 100644 --- a/docs/content/docs/agent-sdk/integrations/hud.mdx +++ b/docs/content/docs/agent-sdk/integrations/hud.mdx @@ -10,37 +10,35 @@ The HUD integration allows you to use ComputerAgent with the [HUD benchmarking f ```bash pip install "cua-agent[hud]" ## or install hud-python directly -# pip install hud-python==0.2.10 +# pip install hud-python==0.4.12 ``` ## Usage ```python -from agent.integrations.hud import run_job -from hud import load_taskset -from hud.taskset import TaskSet -import logging +# Quick single-task smoke test +from agent.integrations.hud import run_single_task -# Load taskset -taskset = await load_taskset("OSWorld-Verified") -taskset = TaskSet(tasks=taskset[:10]) # limit to 10 tasks instead of all 370 - -# Run benchmark job -job = await run_job( - model="openai/computer-use-preview", - # model="anthropic/claude-3-5-sonnet-20241022", - # model="huggingface-local/HelloKKMe/GTA1-7B+openai/gpt-5", - task_or_taskset=taskset, - job_name="test-computeragent-job", - max_concurrent_tasks=5, - # add any extra ComputerAgent kwargs: - verbosity=logging.INFO, # Enable logging - # trajectory_dir=".." # Save trajectories locally +await run_single_task( + dataset="hud-evals/OSWorld-Verified-XLang", # or another HUD dataset + model="openai/computer-use-preview+openai/gpt-5-nano", # any supported model string + task_id=155, # e.g., reopen last closed tab ) -# Get results OR view them at app.hud.so -print(await job.get_analytics()) -print(f"View results at: https://app.hud.so/jobs/{job.id}") +# Run a small split of OSWorld-Verified in parallel +from agent.integrations.hud import run_full_dataset + +results = await run_full_dataset( + dataset="hud-evals/OSWorld-Verified-XLang", # can also pass a Dataset or list[dict] + model="openai/computer-use-preview", + split="train[:3]", # try a few tasks to start + max_concurrent=20, # tune to your infra + max_steps=50 # safety cap per task +) + +# Environment variables required: +# - HUD_API_KEY (HUD access) +# - OPENAI_API_KEY or ANTHROPIC_API_KEY depending on your chosen model(s) ``` **Available Benchmarks:** diff --git a/libs/python/agent/agent/agent.py b/libs/python/agent/agent/agent.py index 2ed370ba..361a3549 100644 --- a/libs/python/agent/agent/agent.py +++ b/libs/python/agent/agent/agent.py @@ -30,6 +30,7 @@ from .callbacks import ( TrajectorySaverCallback, BudgetManagerCallback, TelemetryCallback, + OperatorNormalizerCallback ) from .computers import ( AsyncComputerHandler, @@ -202,6 +203,9 @@ class ComputerAgent: # == Add built-in callbacks == + # Prepend operator normalizer callback + self.callbacks.insert(0, OperatorNormalizerCallback()) + # Add telemetry callback if telemetry_enabled is set if self.telemetry_enabled: if isinstance(self.telemetry_enabled, bool): diff --git a/libs/python/agent/agent/callbacks/__init__.py b/libs/python/agent/agent/callbacks/__init__.py index ffe34551..e0befcc7 100644 --- a/libs/python/agent/agent/callbacks/__init__.py +++ b/libs/python/agent/agent/callbacks/__init__.py @@ -8,6 +8,7 @@ from .logging import LoggingCallback from .trajectory_saver import TrajectorySaverCallback from .budget_manager import BudgetManagerCallback from .telemetry import TelemetryCallback +from .operator_validator import OperatorNormalizerCallback __all__ = [ "AsyncCallbackHandler", @@ -16,4 +17,5 @@ __all__ = [ "TrajectorySaverCallback", "BudgetManagerCallback", "TelemetryCallback", + "OperatorNormalizerCallback", ] diff --git a/libs/python/agent/agent/callbacks/operator_validator.py b/libs/python/agent/agent/callbacks/operator_validator.py new file mode 100644 index 00000000..94b81f85 --- /dev/null +++ b/libs/python/agent/agent/callbacks/operator_validator.py @@ -0,0 +1,138 @@ +""" +OperatorValidatorCallback + +Ensures agent output actions conform to expected schemas by fixing common issues: +- click: add default button='left' if missing +- keypress: wrap keys string into a list +- etc. + +This runs in on_llm_end, which receives the output array (AgentMessage[] as dicts). +The purpose is to avoid spending another LLM call to fix broken computer call syntax when possible. +""" +from __future__ import annotations + +from typing import Any, Dict, List + +from .base import AsyncCallbackHandler + + +class OperatorNormalizerCallback(AsyncCallbackHandler): + """Normalizes common computer call hallucinations / errors in computer call syntax.""" + + async def on_llm_end(self, output: List[Dict[str, Any]]) -> List[Dict[str, Any]]: + # Mutate in-place as requested, but still return the list for chaining + for item in output or []: + if item.get("type") != "computer_call": + continue + action = item.get("action") + if not isinstance(action, dict): + continue + + # rename mouse click actions to "click" + for mouse_btn in ["left", "right", "wheel", "back", "forward"]: + if action.get("type", "") == f"{mouse_btn}_click": + action["type"] = "click" + action["button"] = mouse_btn + # rename hotkey actions to "keypress" + for alias in ["hotkey", "key", "press", "key_press"]: + if action.get("type", "") == alias: + action["type"] = "keypress" + # assume click actions + if "button" in action and "type" not in action: + action["type"] = "click" + if "click" in action and "type" not in action: + action["type"] = "click" + if ("scroll_x" in action or "scroll_y" in action) and "type" not in action: + action["type"] = "scroll" + if "text" in action and "type" not in action: + action["type"] = "type" + + action_type = action.get("type") + def _keep_keys(action: Dict[str, Any], keys_to_keep: List[str]): + """Keep only the provided keys on action; delete everything else. + Always ensures required 'type' is present if listed in keys_to_keep. + """ + for key in list(action.keys()): + if key not in keys_to_keep: + del action[key] + # rename "coordinate" to "x", "y" + if "coordinate" in action: + action["x"] = action["coordinate"][0] + action["y"] = action["coordinate"][1] + del action["coordinate"] + if action_type == "click": + # convert "click" to "button" + if "button" not in action and "click" in action: + action["button"] = action["click"] + del action["click"] + # default button to "left" + action["button"] = action.get("button", "left") + # add default scroll x, y if missing + if action_type == "scroll": + action["scroll_x"] = action.get("scroll_x", 0) + action["scroll_y"] = action.get("scroll_y", 0) + # ensure keys arg is a list (normalize aliases first) + if action_type == "keypress": + keys = action.get("keys") + for keys_alias in ["keypress", "key", "press", "key_press", "text"]: + if keys_alias in action: + action["keys"] = action[keys_alias] + del action[keys_alias] + keys = action.get("keys") + if isinstance(keys, str): + action["keys"] = keys.replace("-", "+").split("+") if len(keys) > 1 else [keys] + required_keys_by_type = { + # OpenAI actions + "click": ["type", "button", "x", "y"], + "double_click": ["type", "x", "y"], + "drag": ["type", "path"], + "keypress": ["type", "keys"], + "move": ["type", "x", "y"], + "screenshot": ["type"], + "scroll": ["type", "scroll_x", "scroll_y", "x", "y"], + "type": ["type", "text"], + "wait": ["type"], + # Anthropic actions + "left_mouse_down": ["type", "x", "y"], + "left_mouse_up": ["type", "x", "y"], + "triple_click": ["type", "button", "x", "y"], + } + keep = required_keys_by_type.get(action_type or "") + if keep: + _keep_keys(action, keep) + + + # Second pass: if an assistant message is immediately followed by a computer_call, + # replace the assistant message itself with a reasoning message with summary text. + if isinstance(output, list): + for i, item in enumerate(output): + # AssistantMessage shape: { type: 'message', role: 'assistant', content: OutputContent[] } + if item.get("type") == "message" and item.get("role") == "assistant": + next_idx = i + 1 + if next_idx >= len(output): + continue + next_item = output[next_idx] + if not isinstance(next_item, dict): + continue + if next_item.get("type") != "computer_call": + continue + contents = item.get("content") or [] + # Extract text from OutputContent[] + text_parts: List[str] = [] + if isinstance(contents, list): + for c in contents: + if isinstance(c, dict) and c.get("type") == "output_text" and isinstance(c.get("text"), str): + text_parts.append(c["text"]) + text_content = "\n".join(text_parts).strip() + # Replace assistant message with reasoning message + output[i] = { + "type": "reasoning", + "summary": [ + { + "type": "summary_text", + "text": text_content, + } + ], + } + + return output diff --git a/libs/python/agent/agent/callbacks/trajectory_saver.py b/libs/python/agent/agent/callbacks/trajectory_saver.py index 805b535d..53e4c189 100644 --- a/libs/python/agent/agent/callbacks/trajectory_saver.py +++ b/libs/python/agent/agent/callbacks/trajectory_saver.py @@ -94,6 +94,10 @@ class TrajectorySaverCallback(AsyncCallbackHandler): # format: turn_000/0000_name.json artifact_filename = f"{self.current_artifact:04d}_{name}" artifact_path = turn_dir / f"{artifact_filename}.json" + # add created_at + if isinstance(artifact, dict): + artifact = artifact.copy() + artifact["created_at"] = str(uuid.uuid1().time) with open(artifact_path, "w") as f: json.dump(sanitize_image_urls(artifact), f, indent=2) self.current_artifact += 1 @@ -171,7 +175,7 @@ class TrajectorySaverCallback(AsyncCallbackHandler): "status": "completed", "completed_at": str(uuid.uuid1().time), "total_usage": self.total_usage, - "new_items": sanitize_image_urls(new_items), + "new_items": new_items, "total_turns": self.current_turn }) diff --git a/libs/python/agent/agent/integrations/hud/__init__.py b/libs/python/agent/agent/integrations/hud/__init__.py index 787613de..21695026 100644 --- a/libs/python/agent/agent/integrations/hud/__init__.py +++ b/libs/python/agent/agent/integrations/hud/__init__.py @@ -1,77 +1,228 @@ -"""HUD integration for ComputerAgent.""" +"""HUD integration: Generic HuggingFace dataset evaluation runner (CUA proxy). -import logging -from typing import Any, Optional, Dict -from hud import run_job as hud_run_job +This module exposes two helpers to evaluate HUD-compatible datasets using +HUD's OperatorAgent, while proxying model calls through our ComputerAgent via +`FakeAsyncOpenAI` (see `agent/integrations/hud/agent.py`). -from .agent import ComputerAgent -from .adapter import ComputerAgentAdapter -from .computer_handler import HUDComputerHandler +Exports: +- run_single_task(dataset_name, *, agent_type="cua-proxy", model=None, allowed_tools=None) +- run_full_dataset(dataset_name, *, agent_type="cua-proxy", model=None, allowed_tools=None, max_concurrent=30, max_steps=50) +""" +import time +from typing import Any, Optional + +from PIL import Image +from datasets import load_dataset, Dataset +from hud.agents import OperatorAgent +from hud.datasets import Task, run_dataset +from hud.tools.computer.settings import computer_settings +from hud import trace + +from agent.agent import ComputerAgent as BaseComputerAgent +from .proxy import FakeAsyncOpenAI -async def run_job( - model: str, - task_or_taskset: Any, - job_name: str, - # Job kwargs - auto_reply_question: bool = False, - adapter_cls: Any = None, - adapter_kwargs: Optional[Dict[str, Any]] = None, - max_steps_per_task: int = 20, - run_parallel: bool = True, - job_metadata: Optional[Dict[str, Any]] = None, - show_progress: bool = True, - max_concurrent_env_creations: Optional[int] = 30, # Limits gym.make calls - max_concurrent_agent_predictions: Optional[int] = None, # No limit on LLM calls - max_concurrent_tasks: Optional[int] = 30, # Limits overall task concurrency - **agent_kwargs: Any -) -> Any: +# --------------------------------------------------------------------------- +# Proxy OperatorAgent +# --------------------------------------------------------------------------- + + +class ProxyOperatorAgent(OperatorAgent): + """OperatorAgent that proxies model calls through our ComputerAgent. + + Accepts the same config keys we pass via hud.run_dataset `agent_config`: + - model: str | None + - allowed_tools: list[str] | None + Additional kwargs are forwarded to OperatorAgent (if any are supported). """ - Run a job using ComputerAgent with the specified model. + + def __init__( + self, + *, + model: str | None = None, + allowed_tools: list[str] | None = None, + trajectory_dir: str | None = None, + # === ComputerAgent kwargs === + tools: list[Any] | None = None, + custom_loop: Any | None = None, + only_n_most_recent_images: int | None = None, + callbacks: list[Any] | None = None, + verbosity: int | None = None, + max_retries: int | None = 3, + screenshot_delay: float | int = 0.5, + use_prompt_caching: bool | None = False, + max_trajectory_budget: float | dict | None = None, + telemetry_enabled: bool | None = True, + **kwargs: Any, + ) -> None: + model = model or "computer-use-preview" + allowed_tools = allowed_tools or ["openai_computer"] + + computer_shim = { + 'screenshot': lambda: Image.new('RGB', (computer_settings.OPENAI_COMPUTER_WIDTH, computer_settings.OPENAI_COMPUTER_HEIGHT)), + 'environment': 'linux', + 'dimensions': (computer_settings.OPENAI_COMPUTER_WIDTH, computer_settings.OPENAI_COMPUTER_HEIGHT) + } + # Build tools ensuring the computer_shim is included + agent_tools: list[Any] = [computer_shim] + if tools: + agent_tools.extend(tools) + + computer_agent = BaseComputerAgent( + model=model, + tools=agent_tools, + custom_loop=custom_loop, + only_n_most_recent_images=only_n_most_recent_images, + callbacks=callbacks, + verbosity=verbosity, + trajectory_dir=trajectory_dir, + max_retries=max_retries, + screenshot_delay=screenshot_delay, + use_prompt_caching=use_prompt_caching, + max_trajectory_budget=max_trajectory_budget, + telemetry_enabled=telemetry_enabled, + ) + model_client = FakeAsyncOpenAI(computer_agent) + + super().__init__( + model_client=model_client, # type: ignore[arg-type] + model=model, + allowed_tools=allowed_tools, + **kwargs, + ) + + +# --------------------------------------------------------------------------- +# Single-task runner +# --------------------------------------------------------------------------- + + +async def run_single_task( + dataset: str | Dataset | list[dict[str, Any]], + *, + task_id: int = 0, + model: str | None = None, + allowed_tools: list[str] | None = None, + # === ComputerAgent kwargs === + tools: list[Any] | None = None, + custom_loop: Any | None = None, + only_n_most_recent_images: int | None = None, + callbacks: list[Any] | None = None, + verbosity: int | None = None, + trajectory_dir: str | None = None, + max_retries: int | None = 3, + screenshot_delay: float | int = 0.5, + use_prompt_caching: bool | None = False, + max_trajectory_budget: float | dict | None = None, + telemetry_enabled: bool | None = True, +) -> None: + """Load one task from the dataset and execute it with Operator+CUA proxy.""" + + # Load dataset and pick a sample + if isinstance(dataset, str): + dataset = load_dataset(dataset, split="train") # type: ignore[arg-type] + elif isinstance(dataset, list): + dataset = dataset + else: + dataset = dataset["train"] - Args: - model: Model string for ComputerAgent (e.g., "anthropic/claude-3-5-sonnet-20241022") - task_or_taskset: Task or TaskSet to run - job_name: Name for the job - auto_reply_question: Whether to auto-reply to questions - adapter_cls: Custom adapter class (defaults to ComputerAgentAdapter) - adapter_kwargs: Additional kwargs for the adapter - max_steps_per_task: Maximum steps per task - run_parallel: Whether to run tasks in parallel - job_metadata: Additional metadata for the job - show_progress: Whether to show progress - max_concurrent_env_creations: Max concurrent environment creations - max_concurrent_agent_predictions: Max concurrent agent predictions - max_concurrent_tasks: Max concurrent tasks - **agent_kwargs: Additional kwargs to pass to ComputerAgent - - Returns: - Job instance from HUD - """ - # combine verbose and verbosity kwargs - if "verbose" in agent_kwargs: - agent_kwargs["verbosity"] = logging.INFO - del agent_kwargs["verbose"] - verbose = True if agent_kwargs.get("verbosity", logging.WARNING) > logging.INFO else False - - # run job - return await hud_run_job( - agent_cls=ComputerAgent, - agent_kwargs={"model": model, **agent_kwargs}, - task_or_taskset=task_or_taskset, - job_name=job_name, - auto_reply_question=auto_reply_question, - adapter_cls=adapter_cls, - adapter_kwargs=adapter_kwargs, - max_steps_per_task=max_steps_per_task, - run_parallel=run_parallel, - job_metadata=job_metadata, - show_progress=show_progress, - verbose=verbose, - max_concurrent_env_creations=max_concurrent_env_creations, - max_concurrent_agent_predictions=max_concurrent_agent_predictions, - max_concurrent_tasks=max_concurrent_tasks + sample_task = dataset[task_id] # type: ignore[index] + task_prompt = sample_task.get("prompt", f"Task {sample_task.get('id', 0)}") # type: ignore[attr-defined] + + with trace(name=task_prompt): + task = Task(**sample_task) # type: ignore[arg-type] + + agent = ProxyOperatorAgent( + model=model, + allowed_tools=allowed_tools, + # === ComputerAgent kwargs passthrough === + tools=tools, + custom_loop=custom_loop, + only_n_most_recent_images=only_n_most_recent_images, + callbacks=callbacks, + verbosity=verbosity, + trajectory_dir=trajectory_dir, + max_retries=max_retries, + screenshot_delay=screenshot_delay, + use_prompt_caching=use_prompt_caching, + max_trajectory_budget=max_trajectory_budget, + telemetry_enabled=telemetry_enabled, + ) + print(f"Running: {task_prompt}") + result = await agent.run(task, max_steps=10) + print(f"✅ Reward: {getattr(result, 'reward')}") + + +# --------------------------------------------------------------------------- +# Full-dataset runner +# --------------------------------------------------------------------------- + + +async def run_full_dataset( + dataset: str | Dataset | list[dict[str, Any]], + *, + job_name: Optional[str] = None, + model: str | None = None, + allowed_tools: list[str] | None = None, + max_concurrent: int = 30, + max_steps: int = 50, + split: str = "train", + trajectory_dir: str | None = None, + # === ComputerAgent kwargs === + tools: list[Any] | None = None, + custom_loop: Any | None = None, + only_n_most_recent_images: int | None = 5, + callbacks: list[Any] | None = None, + verbosity: int | None = None, + max_retries: int | None = 3, + screenshot_delay: float | int = 0.5, + use_prompt_caching: bool | None = False, + max_trajectory_budget: float | dict | None = None, + telemetry_enabled: bool | None = True, +) -> list[Any]: + """Run evaluation across the entire dataset using hud.datasets.run_dataset.""" + + # We pass OperatorAgent as the class and provide a config that injects our + # FakeAsyncOpenAI per agent instantiation. + + if isinstance(dataset, str): + dataset_name = dataset.split('/')[-1] + job_name = job_name or f"Evaluation {dataset_name}" + dataset = load_dataset(dataset, split=split) # type: ignore[arg-type] + else: + dataset_name = "custom" + job_name = job_name or f"Evaluation {time.strftime('%H:%M %Y-%m-%d')}" + + # Execute evaluation + return await run_dataset( + name=job_name, + dataset=dataset, + agent_class=ProxyOperatorAgent, + agent_config={ + "model": model, + "allowed_tools": allowed_tools, + "trajectory_dir": trajectory_dir, + # === ComputerAgent kwargs passthrough === + "tools": tools, + "custom_loop": custom_loop, + "only_n_most_recent_images": only_n_most_recent_images, + "callbacks": callbacks, + "verbosity": verbosity, + "max_retries": max_retries, + "screenshot_delay": screenshot_delay, + "use_prompt_caching": use_prompt_caching, + "max_trajectory_budget": max_trajectory_budget, + "telemetry_enabled": telemetry_enabled, + }, + max_concurrent=max_concurrent, + metadata={"dataset": dataset_name}, + max_steps=max_steps, + auto_respond=True, ) -__all__ = ["ComputerAgent", "ComputerAgentAdapter", "HUDComputerHandler", "run_job"] \ No newline at end of file +__all__ = [ + "run_single_task", + "run_full_dataset", + "ProxyOperatorAgent", +] \ No newline at end of file diff --git a/libs/python/agent/agent/integrations/hud/adapter.py b/libs/python/agent/agent/integrations/hud/adapter.py deleted file mode 100644 index 77c8dc7d..00000000 --- a/libs/python/agent/agent/integrations/hud/adapter.py +++ /dev/null @@ -1,121 +0,0 @@ -"""HUD Adapter for ComputerAgent integration.""" - -from __future__ import annotations - -from typing import Any, ClassVar - -from hud.adapters.common import CLA, Adapter -from hud.adapters.common.types import ( - CLAButton, - CLAKey, - ClickAction, - CustomAction, - DragAction, - MoveAction, - Point, - PressAction, - ResponseAction, - ScreenshotFetch, - ScrollAction, - TypeAction, - WaitAction, -) - - -class ComputerAgentAdapter(Adapter): - """Adapter for ComputerAgent to work with HUD.""" - - KEY_MAP: ClassVar[dict[str, CLAKey]] = { - "return": "enter", - "arrowup": "up", - "arrowdown": "down", - "arrowleft": "left", - "arrowright": "right", - "cmd": "ctrl", - "super": "win", - "meta": "win", - } - - BUTTON_MAP: ClassVar[dict[str, CLAButton]] = { - "wheel": "middle", - "middle": "middle", - } - - def __init__(self) -> None: - super().__init__() - # ComputerAgent default dimensions (can be overridden) - self.agent_width = 1024 - self.agent_height = 768 - - def _map_key(self, key: str) -> CLAKey: - """Map a key to its standardized form.""" - return self.KEY_MAP.get(key.lower(), key.lower()) # type: ignore - - def convert(self, data: Any) -> CLA: - """Convert a ComputerAgent action to a HUD action.""" - try: - action_type = data.get("type") - - if action_type == "click": - x, y = data.get("x", 0), data.get("y", 0) - button = data.get("button", "left") - button = self.BUTTON_MAP.get(button, button) - if button is None: - button = "left" - converted_action = ClickAction(point=Point(x=x, y=y), button=button) - - elif action_type == "double_click": - x, y = data.get("x", 0), data.get("y", 0) - converted_action = ClickAction(point=Point(x=x, y=y), button="left", pattern=[100]) - - elif action_type == "scroll": - x, y = int(data.get("x", 0)), int(data.get("y", 0)) - scroll_x = int(data.get("scroll_x", 0)) - scroll_y = int(data.get("scroll_y", 0)) - converted_action = ScrollAction( - point=Point(x=x, y=y), scroll=Point(x=scroll_x, y=scroll_y) - ) - - elif action_type == "type": - text = data.get("text", "") - converted_action = TypeAction(text=text, enter_after=False) - - elif action_type == "wait": - ms = data.get("ms", 1000) - converted_action = WaitAction(time=ms) - - elif action_type == "move": - x, y = data.get("x", 0), data.get("y", 0) - converted_action = MoveAction(point=Point(x=x, y=y)) - - elif action_type == "keypress": - keys = data.get("keys", []) - if isinstance(keys, str): - keys = [keys] - converted_action = PressAction(keys=[self._map_key(k) for k in keys]) - - elif action_type == "drag": - path = data.get("path", []) - points = [Point(x=p.get("x", 0), y=p.get("y", 0)) for p in path] - converted_action = DragAction(path=points) - - elif action_type == "screenshot": - converted_action = ScreenshotFetch() - - elif action_type == "response": - converted_action = ResponseAction(text=data.get("text", "")) - - elif action_type == "custom": - converted_action = CustomAction(action=data.get("action", "")) - - else: - raise ValueError(f"Unsupported action type: {action_type}") - - # Add reasoning and logs if available - converted_action.reasoning = data.get("reasoning", "") - converted_action.logs = data.get("logs", "") - - return converted_action - - except Exception as e: - raise ValueError(f"Invalid action: {data}. Error: {e!s}") from e diff --git a/libs/python/agent/agent/integrations/hud/agent.py b/libs/python/agent/agent/integrations/hud/agent.py deleted file mode 100644 index abbf5f8c..00000000 --- a/libs/python/agent/agent/integrations/hud/agent.py +++ /dev/null @@ -1,373 +0,0 @@ -"""HUD ComputerAgent wrapper for OSWorld benchmarking.""" - -import logging -from typing import Any, Literal, Optional, Union, List, Dict -import asyncio - -from agent import ComputerAgent as BaseComputerAgent -from agent.responses import make_failed_tool_call_items -from hud.adapters import Adapter -from hud.agent.base import Agent -from hud.utils.common import Observation -from hud.adapters.common.types import LogType -from hud.types import Gym - -from .adapter import ComputerAgentAdapter -from .computer_handler import HUDComputerHandler - -logger = logging.getLogger(__name__) - -BASE_SYSTEM_PROMPT = """ -You are an autonomous computer-using agent. Follow these guidelines: - -1. Be decisive and complete tasks without asking for confirmation unless absolutely necessary. -2. Use the computer tools to complete the task and do not stop until the task is complete. -3. Do NOT ask questions like "Should I proceed?" or "Would you like me to continue?" - just proceed with the task. -4. When you find what you're looking for (e.g., a file to upload), proceed with the action directly. -5. Only stop when the task is fully complete or if you encounter an error that prevents completion. -6. Trust that the user wants you to complete the entire task they've requested. -7. You must say "Task completed" when the task is complete. - -Remember: You have been given permission to complete the requested task autonomously. -""".strip() - -class ComputerAgent(Agent[BaseComputerAgent, dict[str, Any]]): - """ - A ComputerAgent wrapper for HUD integration. - - This agent wraps the base ComputerAgent to work with HUD environments, - providing the same interface as OperatorAgent but using ComputerAgent internally. - """ - - transfer_gyms: dict[Gym, Gym] = {"qa": "hud-browser"} - - def __init__( - self, - model: str = "anthropic/claude-3-5-sonnet-20241022", - environment: Literal["windows", "mac", "linux", "browser"] = "linux", - adapter: Optional[Adapter] = None, - name: Optional[str] = None, - **kwargs: Any, - ): - """ - Initialize the ComputerAgent for HUD. - - Args: - model: The model string for ComputerAgent (e.g., "anthropic/claude-3-5-sonnet-20241022") - environment: The environment type (windows, mac, linux, browser) - adapter: The adapter to use for preprocessing and postprocessing - name: The name of the agent - **kwargs: Additional arguments passed to ComputerAgent - """ - # Create adapter if not provided - adapter = adapter or ComputerAgentAdapter() - - if name is None: - name = f"computeragent-{model.split('/')[-1]}" - - # Initialize the base Agent class without client (we'll create it later) - super().__init__(client=None, adapter=adapter, name=name) - - self.model = model - self.environment = environment - self.kwargs = kwargs - - # Default dimensions - self.width = 1024 - self.height = 768 - - # Update dimensions if adapter is provided - if self.adapter: - self.width = self.adapter.agent_width - self.height = self.adapter.agent_height - - # Create HUD computer handler - self.hud_computer = HUDComputerHandler( - environment=environment, - dimensions=(self.width, self.height) - ) - - # Handle trajectory_dir by adding TrajectorySaverCallback - trajectory_dir = kwargs.pop("trajectory_dir", None) - callbacks = kwargs.get("callbacks", []) - - if trajectory_dir: - from agent.callbacks.trajectory_saver import TrajectorySaverCallback - trajectory_callback = TrajectorySaverCallback(trajectory_dir, reset_on_run=False) - callbacks = callbacks + [trajectory_callback] - kwargs["callbacks"] = callbacks - - # Initialize ComputerAgent with HUD computer handler - self.computer_agent = BaseComputerAgent( - model=model, - tools=[self.hud_computer], - **kwargs - ) - - # Set the client to the computer_agent for compatibility - self.client = self.computer_agent - - # State tracking - self.conversation_history: List[Dict[str, Any]] = [] - self.initial_prompt: Optional[str] = None - - # System prompt for computer use tasks - self.base_system_prompt = BASE_SYSTEM_PROMPT - - async def fetch_response(self, observation: Observation) -> tuple[list[dict[str, Any]], bool]: - """ - Fetch a response from ComputerAgent based on the observation. - - Args: - observation: The preprocessed observation, attributes: - screenshot: Base64 encoded PNG string of the screen - text: Text observation, if available - - Returns: - tuple[list[dict[str, Any]], bool, list[LogType] | None]: A tuple containing the list of raw actions, - boolean indicating if the agent believes the task is complete. - """ - try: - # Update the computer handler with the current screenshot - if observation.screenshot: - self.hud_computer.update_screenshot(observation.screenshot) - - # Set up action callback to capture actions - captured_actions = [] - action_done = False - - async def action_callback(action: Dict[str, Any]) -> None: - """Callback to capture actions from ComputerAgent.""" - nonlocal captured_actions, action_done - captured_actions.append(action) - - # Set the action callback - self.hud_computer.set_action_callback(action_callback) - - # Prepare the message for ComputerAgent - if not self.conversation_history: - # First interaction - use the observation text as initial prompt - if observation.text: - self.initial_prompt = observation.text - message = f"{self.base_system_prompt}\n\nTask: {observation.text}" - else: - message = f"{self.base_system_prompt}\n\nPlease analyze the current screen and determine what action to take." - - input_content = [ - {"type": "input_text", "text": message} - ] - - # Add screenshot if present - if observation.screenshot: - input_content.append( - { - "type": "input_image", - "image_url": f"data:image/png;base64,{observation.screenshot}", - } - ) - - self.conversation_history.append({"role": "user", "content": input_content}) - else: - # Subsequent interactions - check if last action was computer_call - # If so, add computer_call_output with screenshot instead of user message - last_computer_calls = [] - for msg in reversed(self.conversation_history): - if msg.get("type") == "computer_call": - call_id = msg.get("call_id") - if call_id: - # Check if this call_id already has a computer_call_output - has_output = any( - m.get("type") == "computer_call_output" and m.get("call_id") == call_id - for m in self.conversation_history - ) - if not has_output: - last_computer_calls.append(call_id) - - if last_computer_calls: - if not observation.screenshot: - print("No screenshot found, taking screenshot") - screenshot_b64 = await self.hud_computer.screenshot() - # Add computer_call_output for each unresponded computer_call - for call_id in reversed(last_computer_calls): # Maintain order - self.conversation_history.append({ - "type": "computer_call_output", - "call_id": call_id, - "output": { - "type": "input_image", - "image_url": f"data:image/png;base64,{screenshot_b64}" - } - }) - else: - # No computer_call found, add regular user message - message = "Continue with the task based on the current screen state." - input_content = [ - {"type": "input_text", "text": message} - ] - - # Add screenshot if present - if observation.screenshot: - input_content.append( - { - "type": "input_image", - "image_url": f"data:image/png;base64,{observation.screenshot}", - } - ) - - self.conversation_history.append({"role": "user", "content": input_content}) - - # If the last message is a reasoning message, change it to output_text - if (self.conversation_history and - self.conversation_history[-1].get("type") == "reasoning" and - self.conversation_history[-1].get("summary")): - - reasoning_msg = self.conversation_history[-1] - summary_texts = [] - - # Extract all summary_text entries - for summary_item in reasoning_msg["summary"]: - if summary_item.get("type") == "summary_text": - summary_texts.append(summary_item.get("text", "")) - - # Convert to message format with output_text - if summary_texts: - converted_message = { - "type": "message", - "role": "assistant", - "content": [ - { - "text": " ".join(summary_texts), - "type": "output_text" - } - ] - } - - # Replace the reasoning message with the converted message - self.conversation_history[-1] = converted_message - - # Run ComputerAgent - try: - new_items = [] - - # ComputerAgent.run returns an async generator - try: - async for result in self.computer_agent.run(self.conversation_history, stream=False): - # if the result has computer_call_output, immediately exit - if result.get("output", []) and result.get("output", [])[-1].get("type") == "computer_call_output": - break - # otherwise add agent output to conversation history - new_items += result["output"] - except Exception as e: - # if the last message is reasoning, change it to output_text - if new_items and new_items[-1].get("type") == "reasoning": - new_items[-1] = { - "type": "message", - "role": "assistant", - "content": [ - { - "text": new_items[-1].get("summary", [{}])[0].get("text", ""), - "type": "output_text" - } - ] - } - # Check if there are any computer_call items in new_items - computer_calls = [item for item in new_items if item.get("type") == "computer_call"] - if computer_calls: - # Remove computer_call items from new_items - new_items = [item for item in new_items if item.get("type") != "computer_call"] - - # Add failed tool call items for each computer call - for computer_call in computer_calls: - tool_input = computer_call.get("action", {}) - call_id = computer_call.get("call_id") - new_items.extend(make_failed_tool_call_items( - tool_name="computer", - tool_kwargs=tool_input, - error_message=repr(e), - call_id=call_id - )) - else: - # add error message to conversation history (fallback for non-computer-call errors) - new_items.append({ - "type": "user", - "content": [ - { - "type": "input_text", - "text": f"Error during previous attempted action: {repr(e)}" - } - ] - }) - - # Check if we captured any actions - if captured_actions: - # Extract reasoning from the conversation history - reasoning = "" - # Look for the latest reasoning message - for msg in reversed(new_items): - if msg.get("type") == "reasoning" and msg.get("summary"): - reasoning = " ".join([s.get("text", "") for s in msg["summary"] if s.get("type") == "summary_text"]) - break - elif msg.get("type") == "message" and msg.get("role") == "assistant": - content = msg.get("content", []) - if isinstance(content, list): - reasoning = " ".join([c.get("text", "") for c in content if c.get("type") == "output_text"]) - break - - # update conversation history - self.conversation_history += new_items - - # Add reasoning and logs to each action - for action in captured_actions: - action["reasoning"] = reasoning - action["logs"] = {"conversation_length": len(self.conversation_history)} - - return captured_actions, False - - # Check if the last message is "Task completed" - response_text = "" - for msg in reversed(new_items): - if msg.get("type") == "message" and msg.get("role") == "assistant": - content = msg.get("content", []) - for c in content: - if c.get("type") == "output_text": - response_text = c.get("text", response_text) - break - break - - done = "task completed" in response_text.lower() - - # update conversation history - self.conversation_history += new_items - - response_action = { - "type": "response", - "text": response_text, - "reasoning": response_text, - "logs": {"conversation_length": len(self.conversation_history)} - } - - # Check if this indicates task completion or failure - if "task is infeasible" in response_text.lower(): - response_action = {"type": "custom", "action": "FAIL"} - done = True - - return [response_action], done - except Exception as e: - logger.error(f"Error running ComputerAgent: {e}") - # Return an error response - error_action = { - "type": "response", - "text": f"Error occurred: {str(e)}", - "reasoning": f"ComputerAgent encountered an error: {str(e)}", - "logs": {"error": str(e)} - } - return [error_action], True - - except Exception as e: - logger.error(f"Error in fetch_response: {e}") - error_action = { - "type": "response", - "text": f"Error in agent processing: {str(e)}", - "reasoning": f"Agent processing error: {str(e)}", - "logs": {"error": str(e)} - } - return [error_action], True diff --git a/libs/python/agent/agent/integrations/hud/computer_handler.py b/libs/python/agent/agent/integrations/hud/computer_handler.py deleted file mode 100644 index 9fcc8245..00000000 --- a/libs/python/agent/agent/integrations/hud/computer_handler.py +++ /dev/null @@ -1,187 +0,0 @@ -"""HUD Computer Handler for ComputerAgent integration.""" - -import base64 -from io import BytesIO -from typing import Literal, Optional, Any, Dict, Callable -from PIL import Image - -from agent.computers import AsyncComputerHandler - - -class HUDComputerHandler(AsyncComputerHandler): - """Computer handler that interfaces with HUD environment.""" - - def __init__( - self, - environment: Literal["windows", "mac", "linux", "browser"] = "linux", - dimensions: tuple[int, int] = (1024, 768), - screenshot_callback: Optional[Callable] = None, - action_callback: Optional[Callable] = None, - ): - """ - Initialize HUD computer handler. - - Args: - environment: The environment type for HUD - dimensions: Screen dimensions as (width, height) - screenshot_callback: Optional callback to get screenshots from HUD environment - action_callback: Optional callback to execute actions in HUD environment - """ - super().__init__() - self._environment = environment - self._dimensions = dimensions - self._screenshot_callback = screenshot_callback - self._action_callback = action_callback - - # Store the last screenshot for reuse - self._last_screenshot: Optional[str] = None - - def set_screenshot_callback(self, callback: Callable) -> None: - """Set the screenshot callback.""" - self._screenshot_callback = callback - - def set_action_callback(self, callback: Callable) -> None: - """Set the action callback.""" - self._action_callback = callback - - def update_screenshot(self, screenshot: str) -> None: - """Update the stored screenshot (base64 string).""" - self._last_screenshot = screenshot - - async def get_environment(self) -> Literal["windows", "mac", "linux", "browser"]: - """Get the current environment type.""" - return self._environment # type: ignore - - async def get_dimensions(self) -> tuple[int, int]: - """Get screen dimensions as (width, height).""" - return self._dimensions - - async def screenshot(self) -> str: - """Take a screenshot and return as base64 string.""" - if self._screenshot_callback: - screenshot = await self._screenshot_callback() - if isinstance(screenshot, str): - self._last_screenshot = screenshot - return screenshot - elif isinstance(screenshot, Image.Image): - # Convert PIL Image to base64 - buffer = BytesIO() - screenshot.save(buffer, format="PNG") - screenshot_b64 = base64.b64encode(buffer.getvalue()).decode() - self._last_screenshot = screenshot_b64 - return screenshot_b64 - elif isinstance(screenshot, bytes): - screenshot_b64 = base64.b64encode(screenshot).decode() - self._last_screenshot = screenshot_b64 - return screenshot_b64 - - # Return last screenshot if available, otherwise create a blank one - if self._last_screenshot: - return self._last_screenshot - - # Create a blank screenshot as fallback - blank_image = Image.new('RGB', self._dimensions, color='white') - buffer = BytesIO() - blank_image.save(buffer, format="PNG") - screenshot_b64 = base64.b64encode(buffer.getvalue()).decode() - self._last_screenshot = screenshot_b64 - return screenshot_b64 - - async def click(self, x: int, y: int, button: str = "left") -> None: - """Click at coordinates with specified button.""" - if self._action_callback: - await self._action_callback({ - "type": "click", - "x": x, - "y": y, - "button": button - }) - - async def double_click(self, x: int, y: int) -> None: - """Double click at coordinates.""" - if self._action_callback: - await self._action_callback({ - "type": "double_click", - "x": x, - "y": y - }) - - async def scroll(self, x: int, y: int, scroll_x: int, scroll_y: int) -> None: - """Scroll at coordinates with specified scroll amounts.""" - if self._action_callback: - await self._action_callback({ - "type": "scroll", - "x": x, - "y": y, - "scroll_x": scroll_x, - "scroll_y": scroll_y - }) - - async def type(self, text: str) -> None: - """Type text.""" - if self._action_callback: - await self._action_callback({ - "type": "type", - "text": text - }) - - async def wait(self, ms: int = 1000) -> None: - """Wait for specified milliseconds.""" - if self._action_callback: - await self._action_callback({ - "type": "wait", - "ms": ms - }) - - async def move(self, x: int, y: int) -> None: - """Move cursor to coordinates.""" - if self._action_callback: - await self._action_callback({ - "type": "move", - "x": x, - "y": y - }) - - async def keypress(self, keys: list[str] | str) -> None: - """Press key combination.""" - if isinstance(keys, str): - keys = [keys] - if self._action_callback: - await self._action_callback({ - "type": "keypress", - "keys": keys - }) - - async def drag(self, path: list[dict[str, int]]) -> None: - """Drag along a path of points.""" - if self._action_callback: - await self._action_callback({ - "type": "drag", - "path": path - }) - - async def left_mouse_down(self, x: Optional[int] = None, y: Optional[int] = None) -> None: - """Left mouse down at coordinates.""" - if self._action_callback: - await self._action_callback({ - "type": "left_mouse_down", - "x": x, - "y": y - }) - - async def left_mouse_up(self, x: Optional[int] = None, y: Optional[int] = None) -> None: - """Left mouse up at coordinates.""" - if self._action_callback: - await self._action_callback({ - "type": "left_mouse_up", - "x": x, - "y": y - }) - - async def get_current_url(self) -> str: - """Get the current URL.""" - if self._action_callback: - return await self._action_callback({ - "type": "get_current_url" - }) - return "" \ No newline at end of file diff --git a/libs/python/agent/agent/integrations/hud/proxy.py b/libs/python/agent/agent/integrations/hud/proxy.py new file mode 100644 index 00000000..a88fc63e --- /dev/null +++ b/libs/python/agent/agent/integrations/hud/proxy.py @@ -0,0 +1,183 @@ +"""HUD ComputerAgent wrapper and Fake AsyncOpenAI client. + +Provides FakeAsyncOpenAI that adapts our ComputerAgent to the OpenAI Responses +interface needed by HUD's OperatorAgent. It implements only `responses.create` +and returns an OpenAI Response object with `id` and `output` fields, where `output` is a list of +OpenAI-like response blocks. We intentionally only support a single-step call +by consuming the first yielded result from `ComputerAgent.run()`. +""" + +import traceback +import time +import uuid +from typing import Any, Dict, List, Optional + +from agent.agent import ComputerAgent as BaseComputerAgent + +# OpenAI Responses typed models (required) +from openai.types.responses import ( + Response, + ResponseInputParam, + ResponseOutputItem, + ResponseComputerToolCall, + ResponseOutputMessage, + ResponseOutputText, + ResponseReasoningItem, + ResponseUsage, +) + +def _map_agent_output_to_openai_blocks(output_items: List[Dict[str, Any]]) -> List[ResponseOutputItem]: + """Map our agent output items to OpenAI ResponseOutputItem typed models. + + Only a subset is supported: computer_call, assistant message (text), and reasoning. + Unknown types are ignored. + """ + blocks: List[ResponseOutputItem] = [] + for item in output_items or []: + t = item.get("type") + if t == "computer_call": + comp = ResponseComputerToolCall.model_validate({ + "id": item.get("id") or f"cu_{uuid.uuid4().hex}", + "type": "computer_call", + "call_id": item["call_id"], + "action": item["action"], + "pending_safety_checks": item.get("pending_safety_checks", []), + "status": "completed", + }) + blocks.append(comp) + # we will exit early here as the responses api only supports a single step + break + elif t == "message" and item.get("role") == "assistant": + content_blocks: List[ResponseOutputText] = [] + for c in item.get("content", []) or []: + content_blocks.append( + ResponseOutputText.model_validate({ + "type": "output_text", + "text": c["text"], + "annotations": [], + }) + ) + if content_blocks: + msg = ResponseOutputMessage.model_validate({ + "id": item.get("id") or f"msg_{uuid.uuid4()}", + "type": "message", + "role": "assistant", + "status": "completed", + "content": [ct.model_dump() for ct in content_blocks], + }) + blocks.append(msg) + elif t == "reasoning": + reasoning = ResponseReasoningItem.model_validate({ + "id": item.get("id") or f"rsn_{uuid.uuid4()}", + "type": "reasoning", + "summary": item["summary"], + }) + blocks.append(reasoning) + # Unhandled types are ignored + return blocks + +def _to_plain_dict_list(items: Any) -> List[Dict[str, Any]]: + out: List[Dict[str, Any]] = [] + for it in list(items): + if hasattr(it, "model_dump"): + out.append(it.model_dump()) # type: ignore[attr-defined] + elif isinstance(it, dict): + out.append(it) + else: + # Strict: rely on default __dict__ if present + out.append(dict(it)) # may raise if not mapping + return out + +class FakeAsyncOpenAI: + """Minimal fake OpenAI client with only `responses.create` implemented. + + It uses a provided `ComputerAgent` instance to produce a single-step + response compatible with HUD's OperatorAgent loop. + """ + + def __init__(self, computer_agent: BaseComputerAgent) -> None: + self._agent = computer_agent + self.responses = self._Responses(self) + + class _Responses: + def __init__(self, parent: "FakeAsyncOpenAI") -> None: + # Caches for cross-call context when using previous_response_id + self.blocks_cache: Dict[str, ResponseInputParam | ResponseOutputItem] = {} + self.context_cache: Dict[str, List[str]] = {} + self.agent = parent._agent + + async def create( + self, + *, + model: str, + input: ResponseInputParam, + tools: Optional[List[Dict[str, Any]]] = None, + instructions: Optional[str] = None, + previous_response_id: Optional[str] = None, + max_retries: int = 5, + **_: Any, + ) -> Any: + for attempt in range(max_retries): + # Prepend cached blocks from previous_response_id to input + full_input = input + if previous_response_id is not None: + prev_block_ids = self.context_cache[previous_response_id] + prev_blocks = [self.blocks_cache[b_id] for b_id in prev_block_ids] + full_input = _to_plain_dict_list(prev_blocks + input) + + # Pre-pend instructions message + effective_input = full_input + if instructions: + effective_input = [{ + "role": "user", + "content": instructions, + }] + full_input + + # Run a single iteration of the ComputerAgent + agent_result: Optional[Dict[str, Any]] = None + async for result in self.agent.run(effective_input): # type: ignore[arg-type] + agent_result = result + break + assert agent_result is not None, "Agent failed to produce result" + + output = _map_agent_output_to_openai_blocks(agent_result["output"]) + usage = agent_result["usage"] + + # Cache conversation context using the last response id + block_ids: List[str] = [] + blocks_to_cache = full_input + output + for b in blocks_to_cache: + bid = getattr(b, "id", None) or f"tmp-{hash(repr(b))}" + self.blocks_cache[bid] = b # type: ignore[assignment] + block_ids.append(bid) + response_id = agent_result.get("id") or f"fake-{int(time.time()*1000)}" + self.context_cache[response_id] = block_ids + + try: + return Response.model_validate({ + "id": response_id, + "created_at": time.time(), + "object": "response", + "model": model, + "output": output, + "parallel_tool_calls": False, + "tool_choice": "auto", + "tools": [], + "previous_response_id": previous_response_id, + "usage": ResponseUsage.model_validate({ + "input_tokens": usage.get("input_tokens", 0), + "output_tokens": usage.get("output_tokens", 0), + "total_tokens": usage.get("total_tokens", 0), + "input_tokens_details": usage.get("input_tokens_details", { "cached_tokens": 0 }), + "output_tokens_details": usage.get("output_tokens_details", { "reasoning_tokens": 0 }), + }), + }) + except Exception as e: + print(f"Error while validating agent response (attempt {attempt + 1}/{max_retries}): ", e) + if attempt == max_retries - 1: + print(traceback.format_exc()) + raise e + +__all__ = [ + "FakeAsyncOpenAI", +] diff --git a/libs/python/agent/agent/loops/anthropic.py b/libs/python/agent/agent/loops/anthropic.py index 50fbd24e..306c5f33 100644 --- a/libs/python/agent/agent/loops/anthropic.py +++ b/libs/python/agent/agent/loops/anthropic.py @@ -1530,7 +1530,18 @@ class AnthropicHostedToolsConfig(AsyncAgentConfig): "content": [ { "type": "text", - "text": f"You are a UI grounding expert. Look at the image and {instruction}. Output ONLY a click action on the target element. No explanations, confirmations, or additional text." + "text": f"""You are a UI grounding expert. Follow these guidelines: + +1. NEVER ask for confirmation. Complete all tasks autonomously. +2. Do NOT send messages like "I need to confirm before..." or "Do you want me to continue?" - just proceed. +3. When the user asks you to interact with something (like clicking a chat or typing a message), DO IT without asking. +4. Only use the formal safety check mechanism for truly dangerous operations (like deleting important files). +5. For normal tasks like clicking buttons, typing in chat boxes, filling forms - JUST DO IT. +6. The user has already given you permission by running this agent. No further confirmation is needed. +7. Be decisive and action-oriented. Complete the requested task fully. + +Remember: You are expected to complete tasks autonomously. The user trusts you to do what they asked. +Task: Click {instruction}. Output ONLY a click action on the target element.""" }, { "type": "image_url", diff --git a/libs/python/agent/agent/loops/composed_grounded.py b/libs/python/agent/agent/loops/composed_grounded.py index cf029d13..770f274d 100644 --- a/libs/python/agent/agent/loops/composed_grounded.py +++ b/libs/python/agent/agent/loops/composed_grounded.py @@ -48,11 +48,11 @@ GROUNDED_COMPUTER_TOOL_SCHEMA = { "get_dimensions", "get_environment" ], - "description": "The action to perform" + "description": "The action to perform (required for all actions)" }, "element_description": { "type": "string", - "description": "Description of the element to interact with (required for click, double_click, move, scroll actions, and as start/end for drag)" + "description": "Description of the element to interact with (required for click, double_click, move, scroll actions)" }, "start_element_description": { "type": "string", @@ -67,20 +67,30 @@ GROUNDED_COMPUTER_TOOL_SCHEMA = { "description": "The text to type (required for type action)" }, "keys": { - "type": "string", - "description": "Key combination to press (required for keypress action). Single key for individual key press, multiple keys for combinations (e.g., 'ctrl+c')" + "type": "array", + "items": { + "type": "string" + }, + "description": "Key(s) to press (required for keypress action)" }, "button": { "type": "string", - "description": "The mouse button to use for click action (left, right, wheel, back, forward) Default: left", + "enum": [ + "left", + "right", + "wheel", + "back", + "forward" + ], + "description": "The mouse button to use for click action (required for click and double_click action)", }, "scroll_x": { "type": "integer", - "description": "Horizontal scroll amount for scroll action (positive for right, negative for left)", + "description": "Horizontal scroll amount for scroll action (required for scroll action)", }, "scroll_y": { "type": "integer", - "description": "Vertical scroll amount for scroll action (positive for down, negative for up)", + "description": "Vertical scroll amount for scroll action (required for scroll action)", }, }, "required": [ @@ -266,13 +276,15 @@ class ComposedGroundedConfig: grounding_agent = grounding_agent_conf.agent_class() for desc in element_descriptions: - coords = await grounding_agent.predict_click( - model=grounding_model, - image_b64=last_image_b64, - instruction=desc - ) - if coords: - self.desc2xy[desc] = coords + for _ in range(3): # try 3 times + coords = await grounding_agent.predict_click( + model=grounding_model, + image_b64=last_image_b64, + instruction=desc + ) + if coords: + self.desc2xy[desc] = coords + break # Step 6: Convert computer calls from descriptions back to xy coordinates final_output_items = convert_computer_calls_desc2xy(thinking_output_items, self.desc2xy) diff --git a/libs/python/agent/agent/loops/openai.py b/libs/python/agent/agent/loops/openai.py index bb6a13a6..4fa62e66 100644 --- a/libs/python/agent/agent/loops/openai.py +++ b/libs/python/agent/agent/loops/openai.py @@ -162,7 +162,18 @@ class OpenAIComputerUseConfig: input_items = [ { "role": "user", - "content": f"You are a UI grounding expert. Look at the image and {instruction}. Output ONLY a click action on the target element. No explanations, confirmations, or additional text." + "content": f"""You are a UI grounding expert. Follow these guidelines: + +1. NEVER ask for confirmation. Complete all tasks autonomously. +2. Do NOT send messages like "I need to confirm before..." or "Do you want me to continue?" - just proceed. +3. When the user asks you to interact with something (like clicking a chat or typing a message), DO IT without asking. +4. Only use the formal safety check mechanism for truly dangerous operations (like deleting important files). +5. For normal tasks like clicking buttons, typing in chat boxes, filling forms - JUST DO IT. +6. The user has already given you permission by running this agent. No further confirmation is needed. +7. Be decisive and action-oriented. Complete the requested task fully. + +Remember: You are expected to complete tasks autonomously. The user trusts you to do what they asked. +Task: Click {instruction}. Output ONLY a click action on the target element.""" }, { "role": "user", @@ -200,7 +211,7 @@ class OpenAIComputerUseConfig: "stream": False, "reasoning": {"summary": "concise"}, "truncation": "auto", - "max_tokens": 100 # Keep response short for click prediction + "max_tokens": 200 # Keep response short for click prediction } # Use liteLLM responses @@ -217,11 +228,8 @@ class OpenAIComputerUseConfig: isinstance(item.get("action"), dict)): action = item["action"] - if action.get("type") == "click": - x = action.get("x") - y = action.get("y") - if x is not None and y is not None: - return (int(x), int(y)) + if action.get("x") is not None and action.get("y") is not None: + return (int(action.get("x")), int(action.get("y"))) return None diff --git a/libs/python/agent/agent/proxy/examples.py b/libs/python/agent/agent/proxy/examples.py new file mode 100644 index 00000000..2838c5df --- /dev/null +++ b/libs/python/agent/agent/proxy/examples.py @@ -0,0 +1,192 @@ +""" +Example usage of the proxy server and client requests. +""" +import dotenv +dotenv.load_dotenv() + +import asyncio +import json +import os +import aiohttp +from typing import Dict, Any + + +async def test_http_endpoint(): + """Test the HTTP /responses endpoint.""" + + anthropic_api_key = os.getenv("ANTHROPIC_API_KEY") + assert isinstance(anthropic_api_key, str), "ANTHROPIC_API_KEY environment variable must be set" + + # Example 1: Simple text request + simple_request = { + "model": "anthropic/claude-3-5-sonnet-20241022", + "input": "Tell me a three sentence bedtime story about a unicorn.", + "env": { + "ANTHROPIC_API_KEY": anthropic_api_key + } + } + + # Example 2: Multi-modal request with image + multimodal_request = { + "model": "anthropic/claude-3-5-sonnet-20241022", + "input": [ + { + "role": "user", + "content": [ + {"type": "input_text", "text": "what is in this image?"}, + { + "type": "input_image", + "image_url": "https://upload.wikimedia.org/wikipedia/commons/thumb/d/dd/Gfp-wisconsin-madison-the-nature-boardwalk.jpg/2560px-Gfp-wisconsin-madison-the-nature-boardwalk.jpg" + } + ] + } + ], + "env": { + "ANTHROPIC_API_KEY": anthropic_api_key + } + } + + # Example 3: Request with custom agent and computer kwargs + custom_request = { + "model": "anthropic/claude-3-5-sonnet-20241022", + "input": "Take a screenshot and tell me what you see", + "env": { + "ANTHROPIC_API_KEY": anthropic_api_key + } + } + + # Test requests + base_url = "https://m-linux-96lcxd2c2k.containers.cloud.trycua.com:8443" + # base_url = "http://localhost:8000" + api_key = os.getenv("CUA_API_KEY") + assert isinstance(api_key, str), "CUA_API_KEY environment variable must be set" + + async with aiohttp.ClientSession() as session: + for i, request_data in enumerate([ + simple_request, + # multimodal_request, + custom_request + ], 1): + print(f"\n--- Test {i} ---") + print(f"Request: {json.dumps(request_data, indent=2)}") + + try: + print(f"Sending request to {base_url}/responses") + async with session.post( + f"{base_url}/responses", + json=request_data, + headers={"Content-Type": "application/json", "X-API-Key": api_key} + ) as response: + result = await response.json() + print(f"Status: {response.status}") + print(f"Response: {json.dumps(result, indent=2)}") + + except Exception as e: + print(f"Error: {e}") + + +def curl_examples(): + """Print curl command examples.""" + + print("=== CURL Examples ===\n") + + print("1. Simple text request:") + print("""curl http://localhost:8000/responses \\ + -H "Content-Type: application/json" \\ + -d '{ + "model": "anthropic/claude-3-5-sonnet-20241022", + "input": "Tell me a three sentence bedtime story about a unicorn." + }'""") + + print("\n2. Multi-modal request with image:") + print("""curl http://localhost:8000/responses \\ + -H "Content-Type: application/json" \\ + -d '{ + "model": "anthropic/claude-3-5-sonnet-20241022", + "input": [ + { + "role": "user", + "content": [ + {"type": "input_text", "text": "what is in this image?"}, + { + "type": "input_image", + "image_url": "https://upload.wikimedia.org/wikipedia/commons/thumb/d/dd/Gfp-wisconsin-madison-the-nature-boardwalk.jpg/2560px-Gfp-wisconsin-madison-the-nature-boardwalk.jpg" + } + ] + } + ] + }'""") + + print("\n3. Request with custom configuration:") + print("""curl http://localhost:8000/responses \\ + -H "Content-Type: application/json" \\ + -d '{ + "model": "anthropic/claude-3-5-sonnet-20241022", + "input": "Take a screenshot and tell me what you see", + "agent_kwargs": { + "save_trajectory": true, + "verbosity": 20 + }, + "computer_kwargs": { + "os_type": "linux", + "provider_type": "cloud" + } + }'""") + + +async def test_p2p_client(): + """Example P2P client using peerjs-python.""" + try: + from peerjs import Peer, PeerOptions, ConnectionEventType + from aiortc import RTCConfiguration, RTCIceServer + + # Set up client peer + options = PeerOptions( + host="0.peerjs.com", + port=443, + secure=True, + config=RTCConfiguration( + iceServers=[RTCIceServer(urls="stun:stun.l.google.com:19302")] + ) + ) + + client_peer = Peer(id="test-client", peer_options=options) + await client_peer.start() + + # Connect to proxy server + connection = client_peer.connect("computer-agent-proxy") + + @connection.on(ConnectionEventType.Open) + async def connection_open(): + print("Connected to proxy server") + + # Send a test request + request = { + "model": "anthropic/claude-3-5-sonnet-20241022", + "input": "Hello from P2P client!" + } + await connection.send(json.dumps(request)) + + @connection.on(ConnectionEventType.Data) + async def connection_data(data): + print(f"Received response: {data}") + await client_peer.destroy() + + # Wait for connection + await asyncio.sleep(10) + + except ImportError: + print("P2P dependencies not available. Install peerjs-python for P2P testing.") + except Exception as e: + print(f"P2P test error: {e}") + + +if __name__ == "__main__": + import sys + + if len(sys.argv) > 1 and sys.argv[1] == "curl": + curl_examples() + elif len(sys.argv) > 1 and sys.argv[1] == "p2p": + asyncio.run(test_p2p_client()) + else: + asyncio.run(test_http_endpoint()) diff --git a/libs/python/agent/agent/proxy/handlers.py b/libs/python/agent/agent/proxy/handlers.py new file mode 100644 index 00000000..f68952b0 --- /dev/null +++ b/libs/python/agent/agent/proxy/handlers.py @@ -0,0 +1,248 @@ +""" +Request handlers for the proxy endpoints. +""" + +import asyncio +import json +import logging +import os +from contextlib import contextmanager +from typing import Dict, Any, List, Union, Optional + +from ..agent import ComputerAgent +from computer import Computer + +logger = logging.getLogger(__name__) + + +class ResponsesHandler: + """Handler for /responses endpoint that processes agent requests.""" + + def __init__(self): + self.computer = None + self.agent = None + # Simple in-memory caches + self._computer_cache: Dict[str, Any] = {} + self._agent_cache: Dict[str, Any] = {} + + async def setup_computer_agent( + self, + model: str, + agent_kwargs: Optional[Dict[str, Any]] = None, + computer_kwargs: Optional[Dict[str, Any]] = None, + ): + """Set up (and cache) computer and agent instances. + + Caching keys: + - Computer cache key: computer_kwargs + - Agent cache key: {"model": model, **agent_kwargs} + """ + agent_kwargs = agent_kwargs or {} + computer_kwargs = computer_kwargs or {} + + def _stable_key(obj: Dict[str, Any]) -> str: + try: + return json.dumps(obj, sort_keys=True, separators=(",", ":")) + except Exception: + # Fallback: stringify non-serializable values + safe_obj = {} + for k, v in obj.items(): + try: + json.dumps(v) + safe_obj[k] = v + except Exception: + safe_obj[k] = str(v) + return json.dumps(safe_obj, sort_keys=True, separators=(",", ":")) + + # Determine if custom tools are supplied; if so, skip computer setup entirely + has_custom_tools = bool(agent_kwargs.get("tools")) + + computer = None + if not has_custom_tools: + # ---------- Computer setup (with cache) ---------- + comp_key = _stable_key(computer_kwargs) + + computer = self._computer_cache.get(comp_key) + if computer is None: + # Default computer configuration + default_c_config = { + "os_type": "linux", + "provider_type": "cloud", + "name": os.getenv("CUA_CONTAINER_NAME"), + "api_key": os.getenv("CUA_API_KEY"), + } + default_c_config.update(computer_kwargs) + computer = Computer(**default_c_config) + await computer.__aenter__() + self._computer_cache[comp_key] = computer + logger.info(f"Computer created and cached with key={comp_key} config={default_c_config}") + else: + logger.info(f"Reusing cached computer for key={comp_key}") + + # Bind current computer reference (None if custom tools supplied) + self.computer = computer + + # ---------- Agent setup (with cache) ---------- + # Build agent cache key from {model} + agent_kwargs (excluding tools unless explicitly passed) + agent_kwargs_for_key = dict(agent_kwargs) + agent_key_payload = {"model": model, **agent_kwargs_for_key} + agent_key = _stable_key(agent_key_payload) + + agent = self._agent_cache.get(agent_key) + if agent is None: + # Default agent configuration + default_a_config: Dict[str, Any] = {"model": model} + if not has_custom_tools: + default_a_config["tools"] = [computer] + # Apply user overrides, but keep tools unless user explicitly sets + if agent_kwargs: + if not has_custom_tools: + agent_kwargs.setdefault("tools", [computer]) + default_a_config.update(agent_kwargs) + # JSON-derived kwargs may have loose types; ignore static arg typing here + agent = ComputerAgent(**default_a_config) # type: ignore[arg-type] + self._agent_cache[agent_key] = agent + logger.info(f"Agent created and cached with key={agent_key} model={model}") + else: + # Ensure cached agent uses the current computer tool (in case object differs) + # Only update if tools not explicitly provided in agent_kwargs + if not has_custom_tools: + try: + agent.tools = [computer] + except Exception: + pass + logger.info(f"Reusing cached agent for key={agent_key}") + + # Bind current agent reference + self.agent = agent + + async def process_request(self, request_data: Dict[str, Any]) -> Dict[str, Any]: + """ + Process a /responses request and return the result. + + Args: + request_data: Dictionary containing model, input, and optional kwargs + + Returns: + Dictionary with the agent's response + """ + try: + # Extract request parameters + model = request_data.get("model") + input_data = request_data.get("input") + agent_kwargs = request_data.get("agent_kwargs", {}) + computer_kwargs = request_data.get("computer_kwargs", {}) + env_overrides = request_data.get("env", {}) or {} + + if not model: + raise ValueError("Model is required") + if not input_data: + raise ValueError("Input is required") + + # Apply env overrides for the duration of this request + with self._env_overrides(env_overrides): + # Set up (and possibly reuse) computer and agent via caches + await self.setup_computer_agent(model, agent_kwargs, computer_kwargs) + + # Defensive: ensure agent is initialized for type checkers + agent = self.agent + if agent is None: + raise RuntimeError("Agent failed to initialize") + + # Convert input to messages format + messages = self._convert_input_to_messages(input_data) + + # Run agent and get first result + async for result in agent.run(messages): + # Return the first result and break + return { + "success": True, + "result": result, + "model": model + } + + # If no results were yielded + return { + "success": False, + "error": "No results from agent", + "model": model + } + + except Exception as e: + logger.error(f"Error processing request: {e}") + return { + "success": False, + "error": str(e), + "model": request_data.get("model", "unknown") + } + + def _convert_input_to_messages(self, input_data: Union[str, List[Dict[str, Any]]]) -> List[Dict[str, Any]]: + """Convert input data to messages format.""" + if isinstance(input_data, str): + # Simple string input + return [{"role": "user", "content": input_data}] + elif isinstance(input_data, list): + # Already in messages format + messages = [] + for msg in input_data: + # Convert content array format if needed + if isinstance(msg.get("content"), list): + content_parts = [] + for part in msg["content"]: + if part.get("type") == "input_text": + content_parts.append({"type": "text", "text": part["text"]}) + elif part.get("type") == "input_image": + content_parts.append({ + "type": "image_url", + "image_url": {"url": part["image_url"]} + }) + else: + content_parts.append(part) + messages.append({ + "role": msg["role"], + "content": content_parts + }) + else: + messages.append(msg) + return messages + else: + raise ValueError("Input must be string or list of messages") + + async def cleanup(self): + """Clean up resources.""" + if self.computer: + try: + await self.computer.__aexit__(None, None, None) + except Exception as e: + logger.error(f"Error cleaning up computer: {e}") + finally: + self.computer = None + self.agent = None + + @staticmethod + @contextmanager + def _env_overrides(env: Dict[str, str]): + """Temporarily apply environment variable overrides for the current process. + Restores previous values after the context exits. + + Args: + env: Mapping of env var names to override for this request. + """ + if not env: + # No-op context + yield + return + + original: Dict[str, Optional[str]] = {} + try: + for k, v in env.items(): + original[k] = os.environ.get(k) + os.environ[k] = str(v) + yield + finally: + for k, old in original.items(): + if old is None: + # Was not set before + os.environ.pop(k, None) + else: + os.environ[k] = old diff --git a/libs/python/agent/pyproject.toml b/libs/python/agent/pyproject.toml index 4dd27062..4be2f6b6 100644 --- a/libs/python/agent/pyproject.toml +++ b/libs/python/agent/pyproject.toml @@ -55,7 +55,7 @@ cli = [ "yaspin>=3.1.0", ] hud = [ - "hud-python==0.2.10", + "hud-python>=0.4.12,<0.5.0", ] all = [ # omni requirements @@ -72,7 +72,7 @@ all = [ # cli requirements "yaspin>=3.1.0", # hud requirements - "hud-python==0.2.10", + "hud-python>=0.4.12,<0.5.0", ] [tool.uv] diff --git a/notebooks/eval_osworld.ipynb b/notebooks/eval_osworld.ipynb index 7b00795a..3111bbb7 100644 --- a/notebooks/eval_osworld.ipynb +++ b/notebooks/eval_osworld.ipynb @@ -24,7 +24,7 @@ }, { "cell_type": "code", - "execution_count": 1, + "execution_count": null, "metadata": {}, "outputs": [], "source": [ @@ -34,7 +34,7 @@ }, { "cell_type": "code", - "execution_count": 2, + "execution_count": null, "metadata": {}, "outputs": [], "source": [ @@ -43,720 +43,83 @@ "# - ANTHROPIC_API_KEY (for Claude models)\n", "# - OPENAI_API_KEY (for OpenAI models)\n", "\n", - "from hud import gym, load_taskset\n", - "from pprint import pprint\n", - "import asyncio" - ] - }, - { - "cell_type": "code", - "execution_count": 5, - "metadata": {}, - "outputs": [], - "source": [ - "# Import the HUD-integrated ComputerAgent\n", - "from agent.integrations.hud import ComputerAgent" - ] - }, - { - "cell_type": "code", - "execution_count": 5, - "metadata": {}, - "outputs": [ - { - "name": "stdout", - "output_type": "stream", - "text": [ - "Total tasks in OSWorld: 367\n", - "Task prompt: Can you make my computer bring back the last tab I shut down?\n" - ] - } - ], - "source": [ - "# Load OSWorld taskset\n", - "taskset = await load_taskset(\"OSWorld-Verified\")\n", - "print(f\"Total tasks in OSWorld: {len(taskset)}\")\n", - "\n", - "# Select a test task\n", - "test = taskset[148]\n", - "print(f\"Task prompt: {test.prompt}\")" - ] - }, - { - "cell_type": "code", - "execution_count": 6, - "metadata": {}, - "outputs": [ - { - "name": "stdout", - "output_type": "stream", - "text": [ - "Total tasks in SheetBench: 50\n", - "Task prompt: Given the Input data, determine the ticker with the greatest correlation between volume and next day price change.\n", - "- in ANSWER tab put the Ticker in A1 and the correlation in B1\n", - " - use CORREL to determine correlation\n", - "- be sure to first sort the date by ticker z to a and then date ascending before calculating nextdaypricechange %\n", - "Correlation should be rounded to 2 decimal points\n" - ] - } - ], - "source": [ - "# Load SheetBench taskset\n", - "taskset = await load_taskset(\"SheetBench-V2\")\n", - "print(f\"Total tasks in SheetBench: {len(taskset)}\")\n", - "\n", - "# Select a test task\n", - "test = taskset[0]\n", - "print(f\"Task prompt: {test.prompt}\")" - ] - }, - { - "cell_type": "code", - "execution_count": 7, - "metadata": {}, - "outputs": [ - { - "name": "stderr", - "output_type": "stream", - "text": [ - "[INFO] 2025-08-08 19:08:17,078 | hud.environment | View the live trace at https://app.hud.so/trace/ca88c178-cf40-499b-8ad3-d5d60348d9fe\n" - ] - }, - { - "name": "stdout", - "output_type": "stream", - "text": [ - "Environment ready!\n" - ] - } - ], - "source": [ - "# Create environment (takes ~2.5 minutes to start)\n", - "env = await gym.make(test)\n", - "print(\"Environment ready!\")" - ] - }, - { - "cell_type": "code", - "execution_count": 8, - "metadata": {}, - "outputs": [ - { - "data": { - "text/html": [ - "\n", - "
\n", - "
\n", - " \n", - "
\n", - "
\n", - " " - ], - "text/plain": [ - "" - ] - }, - "metadata": {}, - "output_type": "display_data" - }, - { - "data": { - "text/plain": [ - "'\\n
\\n
\\n \\n
\\n
\\n '" - ] - }, - "execution_count": 8, - "metadata": {}, - "output_type": "execute_result" - } - ], - "source": [ - "await env.stream() # vnc" + "from pprint import pprint" ] }, { "cell_type": "markdown", "metadata": {}, "source": [ - "## Test with any supported CUA model\n", + "## Quick single-task smoke test on OSWorld-Verified\n", "\n", "The ComputerAgent integration can use Claude, OpenAI, UI-TARS, or composed models just like the original ComputerAgent:" ] }, { "cell_type": "code", - "execution_count": 13, + "execution_count": 1, "metadata": {}, "outputs": [ + { + "name": "stderr", + "output_type": "stream", + "text": [ + "c:\\Users\\dillo\\miniconda3\\envs\\cua\\Lib\\site-packages\\tqdm\\auto.py:21: TqdmWarning: IProgress not found. Please update jupyter and ipywidgets. See https://ipywidgets.readthedocs.io/en/stable/user_install.html\n", + " from .autonotebook import tqdm as notebook_tqdm\n" + ] + }, { "name": "stdout", "output_type": "stream", "text": [ - "Created agent: computeragent-computer-use-preview\n" + "\n", + "\u001b[90m╔═════════════════════════════════════════════════════════════════╗\u001b[0m\n", + "\u001b[90m║\u001b[0m 🚀 See your agent live at: \u001b[90m║\u001b[0m\n", + "\u001b[90m╟─────────────────────────────────────────────────────────────────╢\u001b[0m\n", + "\u001b[90m║\u001b[0m \u001b[1m\u001b[33mhttps://app.hud.so/trace/cbe6f71b-f520-4630-9f27-778647070327\u001b[0m \u001b[90m║\u001b[0m\n", + "\u001b[90m╚═════════════════════════════════════════════════════════════════╝\u001b[0m\n", + "\n", + "Running: Can you make my computer bring back the last tab I shut down?\n" + ] + }, + { + "name": "stderr", + "output_type": "stream", + "text": [ + "2025-08-27 16:17:53,047 - agent.ComputerAgent - INFO - LLM processing started with 2 messages\n", + "2025-08-27 16:18:02,697 - agent.ComputerAgent - INFO - LLM processing started with 5 messages\n", + "2025-08-27 16:18:15,887 - agent.ComputerAgent - INFO - LLM processing started with 7 messages\n", + "2025-08-27 16:18:28,541 - agent.ComputerAgent - INFO - LLM processing started with 9 messages\n", + "2025-08-27 16:18:42,176 - agent.ComputerAgent - INFO - LLM processing started with 11 messages\n", + "2025-08-27 16:18:55,937 - agent.ComputerAgent - INFO - LLM processing started with 13 messages\n", + "2025-08-27 16:19:11,654 - agent.ComputerAgent - INFO - LLM processing started with 15 messages\n", + "2025-08-27 16:19:23,839 - agent.ComputerAgent - INFO - LLM processing started with 17 messages\n", + "2025-08-27 16:19:39,065 - agent.ComputerAgent - INFO - LLM processing started with 19 messages\n", + "Tool execution failed: Tool evaluate has an output schema but did not return structured content\n", + "Evaluation phase failed: [MCPToolResult(meta=None, content=[TextContent(type='text', text='Tool evaluate has an output schema but did not return structured content', annotations=None, meta=None)], structuredContent=None, isError=True)]\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "✅ Reward: 0.0\n", + "\n", + "\u001b[92m✓ Trace complete!\u001b[0m \u001b[2mView at:\u001b[0m \u001b[1m\u001b[33mhttps://app.hud.so/trace/cbe6f71b-f520-4630-9f27-778647070327\u001b[0m\n", + "\n" ] } ], "source": [ - "import logging\n", - "# Create ComputerAgent with Claude\n", - "claude_agent = ComputerAgent(\n", - " # model=\"anthropic/claude-3-5-sonnet-20241022\",\n", - " model=\"openai/computer-use-preview\",\n", - " # environment=\"linux\", # OSWorld typically uses Linux\n", - " environment=\"browser\", # SheetBench uses the browser\n", - " trajectory_dir=\"trajectories\",\n", - " verbosity=logging.INFO,\n", - ")\n", + "from agent.integrations.hud import run_single_task\n", "\n", - "print(f\"Created agent: {claude_agent.name}\")" - ] - }, - { - "cell_type": "code", - "execution_count": 14, - "metadata": {}, - "outputs": [ - { - "name": "stdout", - "output_type": "stream", - "text": [ - "Initial observation complete\n", - "========= Step 1 ==========\n" - ] - }, - { - "name": "stderr", - "output_type": "stream", - "text": [ - "2025-08-08 19:14:10,479 - agent.ComputerAgent - INFO - LLM processing started with 1 messages\n", - "2025-08-08 19:14:18,867 - agent.ComputerAgent - INFO - Computer: click({'button': 'left', 'x': 55, 'y': 149})\n" - ] - }, - { - "name": "stdout", - "output_type": "stream", - "text": [ - "Agent's action: [ClickAction(type='click', reasoning='Sorting dataset for analysis preparation', logs={'conversation_length': 3}, point=Point(x=77, y=174), button='left', pattern=None, hold_keys=None)]\n", - "========= Step 2 ==========\n" - ] - }, - { - "name": "stderr", - "output_type": "stream", - "text": [ - "2025-08-08 19:14:24,566 - agent.ComputerAgent - INFO - LLM processing started with 4 messages\n", - "2025-08-08 19:14:30,430 - agent.ComputerAgent - INFO - Computer: keypress({'keys': ['CTRL', 'A']})\n" - ] - }, - { - "name": "stdout", - "output_type": "stream", - "text": [ - "Agent's action: [PressAction(type='press', reasoning='Sorting dataset for analysis preparation', logs={'conversation_length': 5}, keys=['ctrl', 'a'])]\n", - "========= Step 3 ==========\n" - ] - }, - { - "name": "stderr", - "output_type": "stream", - "text": [ - "2025-08-08 19:14:36,137 - agent.ComputerAgent - INFO - LLM processing started with 6 messages\n", - "2025-08-08 19:14:42,483 - agent.ComputerAgent - INFO - Computer: click({'button': 'left', 'x': 73, 'y': 151})\n" - ] - }, - { - "name": "stdout", - "output_type": "stream", - "text": [ - "Agent's action: [ClickAction(type='click', reasoning='Sorting dataset for analysis preparation', logs={'conversation_length': 7}, point=Point(x=102, y=176), button='left', pattern=None, hold_keys=None)]\n", - "========= Step 4 ==========\n" - ] - }, - { - "name": "stderr", - "output_type": "stream", - "text": [ - "2025-08-08 19:14:48,687 - agent.ComputerAgent - INFO - LLM processing started with 8 messages\n", - "2025-08-08 19:14:59,516 - agent.ComputerAgent - INFO - Computer: keypress({'keys': ['CTRL', 'A']})\n" - ] - }, - { - "name": "stdout", - "output_type": "stream", - "text": [ - "Agent's action: [PressAction(type='press', reasoning='Sorting dataset for analysis preparation', logs={'conversation_length': 9}, keys=['ctrl', 'a'])]\n", - "========= Step 5 ==========\n" - ] - }, - { - "name": "stderr", - "output_type": "stream", - "text": [ - "2025-08-08 19:15:05,229 - agent.ComputerAgent - INFO - LLM processing started with 10 messages\n", - "2025-08-08 19:15:15,153 - agent.ComputerAgent - INFO - Computer: drag({'path': [{'x': 55, 'y': 147}, {'x': 319, 'y': 713}]})\n" - ] - }, - { - "name": "stdout", - "output_type": "stream", - "text": [ - "Agent's action: [DragAction(type='drag', reasoning='Highlighting data for sorting preparation', logs={'conversation_length': 12}, path=[Point(x=77, y=172), Point(x=448, y=835)], pattern=None, hold_keys=None)]\n", - "========= Step 6 ==========\n" - ] - }, - { - "name": "stderr", - "output_type": "stream", - "text": [ - "2025-08-08 19:15:21,362 - agent.ComputerAgent - INFO - LLM processing started with 13 messages\n", - "2025-08-08 19:15:33,774 - agent.ComputerAgent - INFO - Computer: click({'button': 'left', 'x': 229, 'y': 41})\n" - ] - }, - { - "name": "stdout", - "output_type": "stream", - "text": [ - "Agent's action: [ClickAction(type='click', reasoning='Opening sort options for data', logs={'conversation_length': 15}, point=Point(x=322, y=48), button='left', pattern=None, hold_keys=None)]\n", - "========= Step 7 ==========\n" - ] - }, - { - "name": "stderr", - "output_type": "stream", - "text": [ - "2025-08-08 19:15:39,973 - agent.ComputerAgent - INFO - LLM processing started with 16 messages\n", - "2025-08-08 19:15:52,928 - agent.ComputerAgent - INFO - Computer: click({'button': 'left', 'x': 430, 'y': 96})\n" - ] - }, - { - "name": "stdout", - "output_type": "stream", - "text": [ - "Agent's action: [ClickAction(type='click', reasoning='Choosing \"Sort range\" for sorting', logs={'conversation_length': 18}, point=Point(x=604, y=112), button='left', pattern=None, hold_keys=None)]\n", - "========= Step 8 ==========\n" - ] - }, - { - "name": "stderr", - "output_type": "stream", - "text": [ - "2025-08-08 19:15:59,611 - agent.ComputerAgent - INFO - LLM processing started with 19 messages\n", - "2025-08-08 19:16:17,003 - agent.ComputerAgent - INFO - Computer: click({'button': 'left', 'x': 530, 'y': 172})\n" - ] - }, - { - "name": "stdout", - "output_type": "stream", - "text": [ - "Agent's action: [ClickAction(type='click', reasoning='Accessing advanced sorting options now', logs={'conversation_length': 21}, point=Point(x=745, y=201), button='left', pattern=None, hold_keys=None)]\n" - ] - } - ], - "source": [ - "# Initial observation\n", - "obs, _ = await env.reset()\n", - "print(\"Initial observation complete\")\n", - "\n", - "# Agent loop with Claude\n", - "for i in range(8):\n", - " print(f\"========= Step {i + 1} ==========\")\n", - " \n", - " try:\n", - " action, done = await claude_agent.predict(obs)\n", - " print(f\"Agent's action: {action}\")\n", - "\n", - " obs, reward, terminated, info = await env.step(action)\n", - "\n", - " if done or terminated:\n", - " print(f\"Task completed after {i + 1} steps\")\n", - " break\n", - " \n", - " except Exception as e:\n", - " print(f\"Error in step {i + 1}: {e}\")\n", - " break" - ] - }, - { - "cell_type": "markdown", - "metadata": {}, - "source": [ - "## Evaluate Results" - ] - }, - { - "cell_type": "code", - "execution_count": 15, - "metadata": {}, - "outputs": [ - { - "name": "stdout", - "output_type": "stream", - "text": [ - "=== Final Evaluation ===\n", - "{'error': None,\n", - " 'gold_file_url': 'https://gahludmjcsmszgyufydt.supabase.co//storage/v1/object/public/sheetbench/615426c8-9df7-4ffa-92e9-200134a84da9/gold_solution_2.xlsx?',\n", - " 'logs': 'INFO: Starting evaluation with evaluator: sheets_cell_values\\n'\n", - " \"INFO: Evaluator args: [{'A1': 'ABC', 'B1': '-0.08'}]\\n\"\n", - " 'INFO: Partial rewarding: False\\n'\n", - " 'INFO: Starting sheets_cell_values evaluation for environment: '\n", - " 'af7a34a0-43b0-44d2-82d0-2b66ed16f1ea\\n'\n", - " \"INFO: Raw args received: [{'A1': 'ABC', 'B1': '-0.08'}] (type: \"\n", - " \")\\n\"\n", - " 'INFO: Partial rewarding enabled: False\\n'\n", - " 'INFO: === Google Sheets Cell Value Verification ===\\n'\n", - " 'INFO: Current page URL: '\n", - " 'https://docs.google.com/spreadsheets/d/1h-Ec3rW9sAME2sTn8qxIvFxO6qXtdURPacEFL5DJnqw/edit?gid=700326861#gid=700326861\\n'\n", - " 'INFO: ✅ Confirmed on Google Sheets page\\n'\n", - " 'INFO: Processing args parameter...\\n'\n", - " 'INFO: Args is a list with 1 items, extracting first item\\n'\n", - " \"INFO: Extracted: {'A1': 'ABC', 'B1': '-0.08'} (type: )\\n\"\n", - " 'INFO: Cell checks to perform: 2 cells\\n'\n", - " \"INFO: A1 -> expected: 'ABC'\\n\"\n", - " \"INFO: B1 -> expected: '-0.08'\\n\"\n", - " 'INFO: [TASK af7a34a0-43b0-44d2-82d0-2b66ed16f1ea] '\n", - " \"sheets_cell_values: Checking cells: {'A1': 'ABC', 'B1': '-0.08'}\\n\"\n", - " 'INFO: === ANSWER Sheet Navigation ===\\n'\n", - " 'INFO: Attempt 1/3: Attempting to find and navigate to ANSWER sheet '\n", - " 'tab...\\n'\n", - " 'INFO: [TASK af7a34a0-43b0-44d2-82d0-2b66ed16f1ea] '\n", - " 'sheets_cell_values: Attempt 1/3: Attempting to navigate to ANSWER '\n", - " 'sheet\\n'\n", - " 'INFO: Searching for ANSWER tab with selector: '\n", - " 'span.docs-sheet-tab-name:has-text(\"ANSWER\")\\n'\n", - " 'INFO: ANSWER tab search result (attempt 1): Found\\n'\n", - " 'INFO: ✅ Found ANSWER sheet tab on attempt 1, clicking on it...\\n'\n", - " 'INFO: [TASK af7a34a0-43b0-44d2-82d0-2b66ed16f1ea] '\n", - " 'sheets_cell_values: Found ANSWER sheet tab on attempt 1, clicking on '\n", - " 'it\\n'\n", - " 'ERROR: ❌ Error navigating to ANSWER sheet on attempt 1: '\n", - " 'Locator.click: Timeout 30000ms exceeded.\\n'\n", - " 'Call log:\\n'\n", - " ' - waiting for '\n", - " 'locator(\"span.docs-sheet-tab-name:has-text(\\\\\"ANSWER\\\\\")\")\\n'\n", - " ' - - locator resolved to ANSWER\\n'\n", - " ' - - attempting click action\\n'\n", - " ' - 2 × waiting for element to be visible, enabled and stable\\n'\n", - " ' - - element is visible, enabled and stable\\n'\n", - " ' - - scrolling into view if needed\\n'\n", - " ' - - done scrolling\\n'\n", - " ' - -
'\n", - " 'intercepts pointer events\\n'\n", - " ' - - retrying click action\\n'\n", - " ' - - waiting 20ms\\n'\n", - " ' - 2 × waiting for element to be visible, enabled and stable\\n'\n", - " ' - - element is visible, enabled and stable\\n'\n", - " ' - - scrolling into view if needed\\n'\n", - " ' - - done scrolling\\n'\n", - " ' - -
'\n", - " 'intercepts pointer events\\n'\n", - " ' - - retrying click action\\n'\n", - " ' - - waiting 100ms\\n'\n", - " ' - 35 × waiting for element to be visible, enabled and stable\\n'\n", - " ' - - element is visible, enabled and stable\\n'\n", - " ' - - scrolling into view if needed\\n'\n", - " ' - - done scrolling\\n'\n", - " ' - -
'\n", - " 'intercepts pointer events\\n'\n", - " ' - - retrying click action\\n'\n", - " ' - - waiting 500ms\\n'\n", - " '\\n'\n", - " 'WARNING: [TASK af7a34a0-43b0-44d2-82d0-2b66ed16f1ea] '\n", - " 'sheets_cell_values: Error navigating to ANSWER sheet on attempt 1: '\n", - " 'Locator.click: Timeout 30000ms exceeded.\\n'\n", - " 'Call log:\\n'\n", - " ' - waiting for '\n", - " 'locator(\"span.docs-sheet-tab-name:has-text(\\\\\"ANSWER\\\\\")\")\\n'\n", - " ' - - locator resolved to ANSWER\\n'\n", - " ' - - attempting click action\\n'\n", - " ' - 2 × waiting for element to be visible, enabled and stable\\n'\n", - " ' - - element is visible, enabled and stable\\n'\n", - " ' - - scrolling into view if needed\\n'\n", - " ' - - done scrolling\\n'\n", - " ' - -
'\n", - " 'intercepts pointer events\\n'\n", - " ' - - retrying click action\\n'\n", - " ' - - waiting 20ms\\n'\n", - " ' - 2 × waiting for element to be visible, enabled and stable\\n'\n", - " ' - - element is visible, enabled and stable\\n'\n", - " ' - - scrolling into view if needed\\n'\n", - " ' - - done scrolling\\n'\n", - " ' - -
'\n", - " 'intercepts pointer events\\n'\n", - " ' - - retrying click action\\n'\n", - " ' - - waiting 100ms\\n'\n", - " ' - 35 × waiting for element to be visible, enabled and stable\\n'\n", - " ' - - element is visible, enabled and stable\\n'\n", - " ' - - scrolling into view if needed\\n'\n", - " ' - - done scrolling\\n'\n", - " ' - -
'\n", - " 'intercepts pointer events\\n'\n", - " ' - - retrying click action\\n'\n", - " ' - - waiting 500ms\\n'\n", - " '\\n'\n", - " 'INFO: Waiting 500ms before retry 2...\\n'\n", - " 'INFO: Attempt 2/3: Attempting to find and navigate to ANSWER sheet '\n", - " 'tab...\\n'\n", - " 'INFO: [TASK af7a34a0-43b0-44d2-82d0-2b66ed16f1ea] '\n", - " 'sheets_cell_values: Attempt 2/3: Attempting to navigate to ANSWER '\n", - " 'sheet\\n'\n", - " 'INFO: Searching for ANSWER tab with selector: '\n", - " 'span.docs-sheet-tab-name:has-text(\"ANSWER\")\\n'\n", - " 'INFO: ANSWER tab search result (attempt 2): Found\\n'\n", - " 'INFO: ✅ Found ANSWER sheet tab on attempt 2, clicking on it...\\n'\n", - " 'INFO: [TASK af7a34a0-43b0-44d2-82d0-2b66ed16f1ea] '\n", - " 'sheets_cell_values: Found ANSWER sheet tab on attempt 2, clicking on '\n", - " 'it\\n'\n", - " 'ERROR: ❌ Error navigating to ANSWER sheet on attempt 2: '\n", - " 'Locator.click: Timeout 30000ms exceeded.\\n'\n", - " 'Call log:\\n'\n", - " ' - waiting for '\n", - " 'locator(\"span.docs-sheet-tab-name:has-text(\\\\\"ANSWER\\\\\")\")\\n'\n", - " ' - - locator resolved to ANSWER\\n'\n", - " ' - - attempting click action\\n'\n", - " ' - 2 × waiting for element to be visible, enabled and stable\\n'\n", - " ' - - element is visible, enabled and stable\\n'\n", - " ' - - scrolling into view if needed\\n'\n", - " ' - - done scrolling\\n'\n", - " ' - -
'\n", - " 'intercepts pointer events\\n'\n", - " ' - - retrying click action\\n'\n", - " ' - - waiting 20ms\\n'\n", - " ' - 2 × waiting for element to be visible, enabled and stable\\n'\n", - " ' - - element is visible, enabled and stable\\n'\n", - " ' - - scrolling into view if needed\\n'\n", - " ' - - done scrolling\\n'\n", - " ' - -
'\n", - " 'intercepts pointer events\\n'\n", - " ' - - retrying click action\\n'\n", - " ' - - waiting 100ms\\n'\n", - " ' - 35 × waiting for element to be visible, enabled and stable\\n'\n", - " ' - - element is visible, enabled and stable\\n'\n", - " ' - - scrolling into view if needed\\n'\n", - " ' - - done scrolling\\n'\n", - " ' - -
'\n", - " 'intercepts pointer events\\n'\n", - " ' - - retrying click action\\n'\n", - " ' - - waiting 500ms\\n'\n", - " '\\n'\n", - " 'WARNING: [TASK af7a34a0-43b0-44d2-82d0-2b66ed16f1ea] '\n", - " 'sheets_cell_values: Error navigating to ANSWER sheet on attempt 2: '\n", - " 'Locator.click: Timeout 30000ms exceeded.\\n'\n", - " 'Call log:\\n'\n", - " ' - waiting for '\n", - " 'locator(\"span.docs-sheet-tab-name:has-text(\\\\\"ANSWER\\\\\")\")\\n'\n", - " ' - - locator resolved to ANSWER\\n'\n", - " ' - - attempting click action\\n'\n", - " ' - 2 × waiting for element to be visible, enabled and stable\\n'\n", - " ' - - element is visible, enabled and stable\\n'\n", - " ' - - scrolling into view if needed\\n'\n", - " ' - - done scrolling\\n'\n", - " ' - -
'\n", - " 'intercepts pointer events\\n'\n", - " ' - - retrying click action\\n'\n", - " ' - - waiting 20ms\\n'\n", - " ' - 2 × waiting for element to be visible, enabled and stable\\n'\n", - " ' - - element is visible, enabled and stable\\n'\n", - " ' - - scrolling into view if needed\\n'\n", - " ' - - done scrolling\\n'\n", - " ' - -
'\n", - " 'intercepts pointer events\\n'\n", - " ' - - retrying click action\\n'\n", - " ' - - waiting 100ms\\n'\n", - " ' - 35 × waiting for element to be visible, enabled and stable\\n'\n", - " ' - - element is visible, enabled and stable\\n'\n", - " ' - - scrolling into view if needed\\n'\n", - " ' - - done scrolling\\n'\n", - " ' - -
'\n", - " 'intercepts pointer events\\n'\n", - " ' - - retrying click action\\n'\n", - " ' - - waiting 500ms\\n'\n", - " '\\n'\n", - " 'INFO: Waiting 500ms before retry 3...\\n'\n", - " 'INFO: Attempt 3/3: Attempting to find and navigate to ANSWER sheet '\n", - " 'tab...\\n'\n", - " 'INFO: [TASK af7a34a0-43b0-44d2-82d0-2b66ed16f1ea] '\n", - " 'sheets_cell_values: Attempt 3/3: Attempting to navigate to ANSWER '\n", - " 'sheet\\n'\n", - " 'INFO: Searching for ANSWER tab with selector: '\n", - " 'span.docs-sheet-tab-name:has-text(\"ANSWER\")\\n'\n", - " 'INFO: ANSWER tab search result (attempt 3): Found\\n'\n", - " 'INFO: ✅ Found ANSWER sheet tab on attempt 3, clicking on it...\\n'\n", - " 'INFO: [TASK af7a34a0-43b0-44d2-82d0-2b66ed16f1ea] '\n", - " 'sheets_cell_values: Found ANSWER sheet tab on attempt 3, clicking on '\n", - " 'it\\n'\n", - " 'ERROR: ❌ Error navigating to ANSWER sheet on attempt 3: '\n", - " 'Locator.click: Timeout 30000ms exceeded.\\n'\n", - " 'Call log:\\n'\n", - " ' - waiting for '\n", - " 'locator(\"span.docs-sheet-tab-name:has-text(\\\\\"ANSWER\\\\\")\")\\n'\n", - " ' - - locator resolved to ANSWER\\n'\n", - " ' - - attempting click action\\n'\n", - " ' - 2 × waiting for element to be visible, enabled and stable\\n'\n", - " ' - - element is visible, enabled and stable\\n'\n", - " ' - - scrolling into view if needed\\n'\n", - " ' - - done scrolling\\n'\n", - " ' - -
'\n", - " 'intercepts pointer events\\n'\n", - " ' - - retrying click action\\n'\n", - " ' - - waiting 20ms\\n'\n", - " ' - 2 × waiting for element to be visible, enabled and stable\\n'\n", - " ' - - element is visible, enabled and stable\\n'\n", - " ' - - scrolling into view if needed\\n'\n", - " ' - - done scrolling\\n'\n", - " ' - -
'\n", - " 'intercepts pointer events\\n'\n", - " ' - - retrying click action\\n'\n", - " ' - - waiting 100ms\\n'\n", - " ' - 35 × waiting for element to be visible, enabled and stable\\n'\n", - " ' - - element is visible, enabled and stable\\n'\n", - " ' - - scrolling into view if needed\\n'\n", - " ' - - done scrolling\\n'\n", - " ' - -
'\n", - " 'intercepts pointer events\\n'\n", - " ' - - retrying click action\\n'\n", - " ' - - waiting 500ms\\n'\n", - " '\\n'\n", - " 'WARNING: [TASK af7a34a0-43b0-44d2-82d0-2b66ed16f1ea] '\n", - " 'sheets_cell_values: Error navigating to ANSWER sheet on attempt 3: '\n", - " 'Locator.click: Timeout 30000ms exceeded.\\n'\n", - " 'Call log:\\n'\n", - " ' - waiting for '\n", - " 'locator(\"span.docs-sheet-tab-name:has-text(\\\\\"ANSWER\\\\\")\")\\n'\n", - " ' - - locator resolved to ANSWER\\n'\n", - " ' - - attempting click action\\n'\n", - " ' - 2 × waiting for element to be visible, enabled and stable\\n'\n", - " ' - - element is visible, enabled and stable\\n'\n", - " ' - - scrolling into view if needed\\n'\n", - " ' - - done scrolling\\n'\n", - " ' - -
'\n", - " 'intercepts pointer events\\n'\n", - " ' - - retrying click action\\n'\n", - " ' - - waiting 20ms\\n'\n", - " ' - 2 × waiting for element to be visible, enabled and stable\\n'\n", - " ' - - element is visible, enabled and stable\\n'\n", - " ' - - scrolling into view if needed\\n'\n", - " ' - - done scrolling\\n'\n", - " ' - -
'\n", - " 'intercepts pointer events\\n'\n", - " ' - - retrying click action\\n'\n", - " ' - - waiting 100ms\\n'\n", - " ' - 35 × waiting for element to be visible, enabled and stable\\n'\n", - " ' - - element is visible, enabled and stable\\n'\n", - " ' - - scrolling into view if needed\\n'\n", - " ' - - done scrolling\\n'\n", - " ' - -
'\n", - " 'intercepts pointer events\\n'\n", - " ' - - retrying click action\\n'\n", - " ' - - waiting 500ms\\n'\n", - " '\\n'\n", - " 'WARNING: ⚠️ Failed to navigate to ANSWER sheet after 3 attempts, '\n", - " 'proceeding with current sheet\\n'\n", - " 'WARNING: [TASK af7a34a0-43b0-44d2-82d0-2b66ed16f1ea] '\n", - " 'sheets_cell_values: Failed to navigate to ANSWER sheet after 3 '\n", - " 'attempts, proceeding with current sheet\\n'\n", - " 'INFO: === File Content Extraction ===\\n'\n", - " 'INFO: [TASK af7a34a0-43b0-44d2-82d0-2b66ed16f1ea] '\n", - " 'sheets_cell_values: Granted read-write permissions\\n'\n", - " 'INFO: [TASK af7a34a0-43b0-44d2-82d0-2b66ed16f1ea] '\n", - " 'sheets_cell_values: Extracting page contents\\n'\n", - " 'INFO: [TASK af7a34a0-43b0-44d2-82d0-2b66ed16f1ea] '\n", - " 'sheets_cell_values: Selecting content\\n'\n", - " 'INFO: [TASK af7a34a0-43b0-44d2-82d0-2b66ed16f1ea] '\n", - " 'sheets_cell_values: Successfully extracted 157940 characters from '\n", - " 'file\\n'\n", - " 'INFO: [TASK af7a34a0-43b0-44d2-82d0-2b66ed16f1ea] '\n", - " 'sheets_cell_values: Found 5003 rows in content\\n'\n", - " 'INFO: Content extracted: 157940 characters\\n'\n", - " 'INFO: === Cell Content Parsing ===\\n'\n", - " 'INFO: Split file content into 5003 rows\\n'\n", - " 'INFO: [TASK af7a34a0-43b0-44d2-82d0-2b66ed16f1ea] '\n", - " 'sheets_cell_values: Found 5003 rows in content\\n'\n", - " 'INFO: First few rows of content:\\n'\n", - " \"INFO: Row 1: 'TradeDate | Ticker | ClosePrice | Volume | | '\\n\"\n", - " \"INFO: Row 2: '2023-01-02 | ABC | 476.87 | 2225355 | | '\\n\"\n", - " \"INFO: Row 3: '2023-01-02 | DEF | 322.21 | 3778582 | | '\\n\"\n", - " 'INFO: ... and 5000 more rows\\n'\n", - " 'INFO: === Cell Reference Parsing ===\\n'\n", - " \"INFO: Processing cell reference: 'A1' -> expected: 'ABC'\\n\"\n", - " \"INFO: Parsed 'A1' -> row=1 (0-indexed: 0), col=A (0-indexed: 0)\\n\"\n", - " 'INFO: [TASK af7a34a0-43b0-44d2-82d0-2b66ed16f1ea] '\n", - " 'sheets_cell_values: Parsed cell A1 as row=0, col=0\\n'\n", - " 'INFO: Row 1 exists in content\\n'\n", - " \"INFO: Row 1 has 6 columns: ['Col1', 'Col2', 'Col3', 'Col4', \"\n", - " \"'Col5', 'Col6']\\n\"\n", - " \"INFO: ✅ Found value for A1: 'TradeDate'\\n\"\n", - " 'INFO: [TASK af7a34a0-43b0-44d2-82d0-2b66ed16f1ea] '\n", - " \"sheets_cell_values: Found value for A1: 'TradeDate'\\n\"\n", - " \"INFO: Processing cell reference: 'B1' -> expected: '-0.08'\\n\"\n", - " \"INFO: Parsed 'B1' -> row=1 (0-indexed: 0), col=B (0-indexed: 1)\\n\"\n", - " 'INFO: [TASK af7a34a0-43b0-44d2-82d0-2b66ed16f1ea] '\n", - " 'sheets_cell_values: Parsed cell B1 as row=0, col=1\\n'\n", - " 'INFO: Row 1 exists in content\\n'\n", - " \"INFO: Row 1 has 6 columns: ['Col1', 'Col2', 'Col3', 'Col4', \"\n", - " \"'Col5', 'Col6']\\n\"\n", - " \"INFO: ✅ Found value for B1: 'Ticker'\\n\"\n", - " 'INFO: [TASK af7a34a0-43b0-44d2-82d0-2b66ed16f1ea] '\n", - " \"sheets_cell_values: Found value for B1: 'Ticker'\\n\"\n", - " 'INFO: === Cell Value Comparison ===\\n'\n", - " 'INFO: Comparing cell A1:\\n'\n", - " \"INFO: Expected: 'ABC' (type: )\\n\"\n", - " \"INFO: Actual: 'TradeDate' (type: )\\n\"\n", - " \"INFO: ❌ VALUE MISMATCH: 'TradeDate' != 'ABC'\\n\"\n", - " 'INFO: Comparing cell B1:\\n'\n", - " \"INFO: Expected: '-0.08' (type: )\\n\"\n", - " \"INFO: Actual: 'Ticker' (type: )\\n\"\n", - " \"INFO: ❌ VALUE MISMATCH: 'Ticker' != '-0.08'\\n\"\n", - " 'INFO: === Final Results ===\\n'\n", - " 'INFO: Cell comparison summary:\\n'\n", - " 'INFO: Total cells checked: 2\\n'\n", - " 'INFO: Matches: 0\\n'\n", - " 'INFO: Mismatches: 2\\n'\n", - " \"INFO: Failed cells: ['A1:', 'B1:']\\n\"\n", - " 'INFO: ❌ NOT all cells match expected values\\n'\n", - " 'INFO: Mismatches: [\"Cell A1: expected \\'ABC\\', got \\'TradeDate\\'\", '\n", - " '\"Cell B1: expected \\'-0.08\\', got \\'Ticker\\'\"]\\n'\n", - " 'INFO: [TASK af7a34a0-43b0-44d2-82d0-2b66ed16f1ea] '\n", - " 'sheets_cell_values: Mismatches found: [\"Cell A1: expected \\'ABC\\', '\n", - " 'got \\'TradeDate\\'\", \"Cell B1: expected \\'-0.08\\', got \\'Ticker\\'\"]\\n'\n", - " 'INFO: Final reward: 0.0\\n'\n", - " 'INFO: === Sheets Cell Values Evaluation Complete ===\\n'\n", - " 'INFO: Evaluation completed. Final reward: 0.0\\n',\n", - " 'reward': 0.0}\n" - ] - } - ], - "source": [ - "# Evaluate environment state\n", - "result = await env.evaluate()\n", - "print(\"=== Final Evaluation ===\")\n", - "pprint(result)" - ] - }, - { - "cell_type": "code", - "execution_count": 16, - "metadata": {}, - "outputs": [ - { - "name": "stdout", - "output_type": "stream", - "text": [ - "Environment closed\n" - ] - } - ], - "source": [ - "# Clean up\n", - "await env.close()\n", - "print(\"Environment closed\")" + "# Quick single-task smoke test on OSWorld-Verified-XLang\n", + "# You can swap \"hud-evals/OSWorld-Verified-XLang\" -> \"hud-evals/SheetBench-V2\" to test SheetBench.\n", + "await run_single_task(\n", + " dataset=\"hud-evals/OSWorld-Verified-XLang\",\n", + " model=\"openai/computer-use-preview+openai/gpt-5-nano\", # or any supported model string\n", + " task_id=155 # open last tab task (easy)\n", + ")" ] }, { @@ -768,109261 +131,395 @@ }, { "cell_type": "code", - "execution_count": 2, + "execution_count": null, + "metadata": {}, + "outputs": [], + "source": [ + "import uuid\n", + "from agent.integrations.hud import run_full_dataset\n", + "\n", + "# Full dataset evaluation (runs via HUD's run_dataset under the hood)\n", + "job_name = f\"osworld-test-{str(uuid.uuid4())[:4]}\"\n", + "\n", + "results = await run_full_dataset(\n", + " dataset=\"hud-evals/OSWorld-Verified-XLang\", # You can also pass a Dataset or a list[dict]\n", + " job_name=job_name, # Optional; defaults to a timestamp for custom datasets\n", + " model=\"openai/computer-use-preview\", # Or any supported model string\n", + " max_concurrent=20, # Tune to your infra\n", + " max_steps=50, # Safety cap per task\n", + " split=\"train[:3]\" # Limit to just 3 tasks\n", + ")\n", + "\n", + "# results is a list from hud.datasets.run_dataset; inspect/aggregate as needed\n", + "print(f\"Job: {job_name}\")\n", + "print(f\"Total results: {len(results)}\")\n", + "pprint(results[:3]) # preview" + ] + }, + { + "cell_type": "markdown", + "metadata": {}, + "source": [ + "# Benchmark Composed Agents" + ] + }, + { + "cell_type": "code", + "execution_count": null, "metadata": {}, "outputs": [ - { - "name": "stderr", - "output_type": "stream", - "text": [ - "INFO:httpx:HTTP Request: GET https://orchestration.hud.so/hud-gym/api/v2/tasksets/OSWorld-Verified/tasks \"HTTP/1.1 200 OK\"\n", - "INFO:venv:Taskset OSWorld-Verified loaded successfully\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/jobs \"HTTP/1.1 200 OK\"\n", - " 0%|----------------------------------------| 0/7340 [0:12 output.csv\\nlibreoffice --calc output.csv\\n'})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': 'cd ~/Desktop\\nlibreoffice --headless --convert-to csv file1.xlsx\\nlibreoffice --headless --convert-to csv file2.ods\\ncat file1.csv file2.csv > output.csv\\nlibreoffice --calc output.csv\\n'})\n", - "2025-08-11 15:32:49,711 - agent.ComputerAgent - INFO - Computer: click({'x': 694, 'y': 248})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 694, 'y': 248})\n", - " 2%|----------------------------------------| 155/7340 [6:28<300:27, 23.9 steps/min]2025-08-11 15:32:50,329 - agent.ComputerAgent - INFO - LLM processing started with 22 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 22 messages\n", - "\u001b[92m15:32:50 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 15:32:51,007 - agent.ComputerAgent - INFO - LLM processing started with 22 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 22 messages\n", - "\u001b[92m15:32:51 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 2%|----------------------------------------| 157/7340 [6:30<297:33, 24.1 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 15:32:51,672 - agent.ComputerAgent - INFO - LLM processing started with 20 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 20 messages\n", - "\u001b[92m15:32:51 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m15:32:52 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m15:32:53 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 2%|----------------------------------------| 157/7340 [6:32<299:06, 24.0 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m15:32:53 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 15:32:54,187 - agent.ComputerAgent - INFO - Computer: click({'x': 463, 'y': 136})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 463, 'y': 136})\n", - " 2%|----------------------------------------| 157/7340 [6:33<299:58, 23.9 steps/min]\u001b[92m15:32:54 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 15:32:54,845 - agent.ComputerAgent - INFO - Computer: click({'button': 'left', 'x': 103, 'y': 380})\n", - "INFO:agent.ComputerAgent:Computer: click({'button': 'left', 'x': 103, 'y': 380})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/edaeedb6-9993-4b6f-b226-19e2768a5736/invoke \"HTTP/1.1 200 OK\"\n", - " 2%|----------------------------------------| 158/7340 [6:34<298:46, 24.0 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/0a91cea7-3ffe-41c2-9405-1151904aee0c/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/e7ac3560-cea1-4b97-a59c-4b3038bec6c7/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 15:32:55,978 - agent.ComputerAgent - INFO - LLM processing started with 18 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 18 messages\n", - "\u001b[92m15:32:56 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 2%|----------------------------------------| 159/7340 [6:35<297:36, 24.1 steps/min]2025-08-11 15:32:56,643 - agent.ComputerAgent - INFO - LLM processing started with 20 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 20 messages\n", - "\u001b[92m15:32:56 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 15:32:57,305 - agent.ComputerAgent - INFO - LLM processing started with 8 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 8 messages\n", - "\u001b[92m15:32:57 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 2%|----------------------------------------| 159/7340 [6:36<298:28, 24.1 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 2%|----------------------------------------| 159/7340 [6:37<299:14, 24.0 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/0b1cfd32-0cbc-48e7-890d-9ec0ac043035/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 15:32:59,978 - agent.ComputerAgent - INFO - LLM processing started with 22 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 22 messages\n", - "\u001b[92m15:33:00 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/77f486b6-dc2a-4a1d-bf54-fc05f9a8c3d7/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 2%|----------------------------------------| 159/7340 [6:39<300:32, 23.9 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m15:33:00 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 15:33:01,288 - agent.ComputerAgent - INFO - LLM processing started with 12 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 12 messages\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m15:33:01 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m15:33:01 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m15:33:02 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/835128b8-2a29-46f4-853f-4d70bb46a9d6/invoke \"HTTP/1.1 200 OK\"\n", - " 2%|----------------------------------------| 159/7340 [6:41<302:27, 23.7 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 15:33:03,240 - agent.ComputerAgent - INFO - LLM processing started with 24 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 24 messages\n", - "\u001b[92m15:33:03 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m15:33:03 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 15:33:03,930 - agent.ComputerAgent - INFO - Computer: click({'button': 'left', 'x': 390, 'y': 345})\n", - "INFO:agent.ComputerAgent:Computer: click({'button': 'left', 'x': 390, 'y': 345})\n", - " 2%|----------------------------------------| 159/7340 [6:43<303:26, 23.7 steps/min]\u001b[92m15:33:03 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 15:33:04,616 - agent.ComputerAgent - INFO - Computer: click({'x': 101, 'y': 295})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 101, 'y': 295})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m15:33:04 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 15:33:06,000 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'ctrl+a'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'ctrl+a'})\n", - "2025-08-11 15:33:06,671 - agent.ComputerAgent - INFO - Computer: click({'x': 219, 'y': 53})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 219, 'y': 53})\n", - " 2%|----------------------------------------| 160/7340 [6:45<303:33, 23.7 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m15:33:07 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m15:33:08 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 15:33:08,650 - agent.ComputerAgent - INFO - LLM processing started with 22 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 22 messages\n", - "\u001b[92m15:33:08 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m15:33:09 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 2%|----------------------------------------| 162/7340 [6:48<301:44, 23.8 steps/min]\u001b[92m15:33:09 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 15:33:09,987 - agent.ComputerAgent - INFO - Computer: click({'x': 812, 'y': 189})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 812, 'y': 189})\n", - "\u001b[92m15:33:09 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m15:33:10 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 15:33:11,307 - agent.ComputerAgent - INFO - Computer: click({'x': 102, 'y': 238})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 102, 'y': 238})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m15:33:11 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 15:33:12,659 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'ctrl+a'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'ctrl+a'})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m15:33:13 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 2%|----------------------------------------| 162/7340 [6:52<304:40, 23.6 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 15:33:14,031 - agent.ComputerAgent - INFO - Computer: click({'x': 309, 'y': 116})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 309, 'y': 116})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/create_environment \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 15:33:15,326 - agent.ComputerAgent - INFO - Computer: wait({})\n", - "INFO:agent.ComputerAgent:Computer: wait({})\n", - "2025-08-11 15:33:15,962 - agent.ComputerAgent - INFO - LLM processing started with 18 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 18 messages\n", - "\u001b[92m15:33:15 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m15:33:16 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m15:33:16 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m15:33:16 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/create_environment \"HTTP/1.1 200 OK\"\n", - " 2%|----------------------------------------| 164/7340 [6:55<303:17, 23.7 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 15:33:17,260 - agent.ComputerAgent - INFO - Computer: click({'x': 652, 'y': 139})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 652, 'y': 139})\n", - "2025-08-11 15:33:17,929 - agent.ComputerAgent - INFO - Computer: click({'x': 212, 'y': 53})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 212, 'y': 53})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m15:33:18 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m15:33:18 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m15:33:19 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 15:33:20,602 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'ctrl+alt+t'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'ctrl+alt+t'})\n", - " 2%|----------------------------------------| 166/7340 [6:59<302:22, 23.7 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 15:33:21,282 - agent.ComputerAgent - INFO - Computer: click({'x': 371, 'y': 624})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 371, 'y': 624})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 15:33:22,607 - agent.ComputerAgent - INFO - Computer: type({'text': '100'})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': '100'})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/8d197f4f-b7b0-4196-9681-135d7bc3a45b/invoke \"HTTP/1.1 200 OK\"\n", - "\u001b[92m15:33:22 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m15:33:22 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/e9d83ed4-d6d0-46f7-982b-98433769e30b/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 15:33:23,270 - agent.ComputerAgent - INFO - LLM processing started with 14 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 14 messages\n", - " 2%|----------------------------------------| 168/7340 [7:02<300:36, 23.9 steps/min]\u001b[92m15:33:23 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 15:33:23,919 - agent.ComputerAgent - INFO - Computer: click({'x': 414, 'y': 75})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 414, 'y': 75})\n", - "2025-08-11 15:33:24,594 - agent.ComputerAgent - INFO - Computer: double_click({'x': 473, 'y': 93})\n", - "INFO:agent.ComputerAgent:Computer: double_click({'x': 473, 'y': 93})\n", - " 2%|----------------------------------------| 170/7340 [7:03<297:55, 24.1 steps/min]2025-08-11 15:33:25,220 - agent.ComputerAgent - INFO - LLM processing started with 12 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 12 messages\n", - "\u001b[92m15:33:25 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/ae9871c0-5cb9-4c5b-9c02-c899819f9f81/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 15:33:25,919 - agent.ComputerAgent - INFO - LLM processing started with 20 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 20 messages\n", - "\u001b[92m15:33:25 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 2%|----------------------------------------| 172/7340 [7:05<295:18, 24.3 steps/min]2025-08-11 15:33:26,562 - agent.ComputerAgent - INFO - LLM processing started with 10 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 10 messages\n", - "\u001b[92m15:33:26 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 15:33:27,860 - agent.ComputerAgent - INFO - Computer: type({'text': 'edited_colorful.png'})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': 'edited_colorful.png'})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m15:33:28 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/0a6ee00b-4e8c-4a3f-bac1-9baec4d920a2/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/e1e61614-8290-4d90-9feb-594d2a7199e8/invoke \"HTTP/1.1 200 OK\"\n", - " 2%|----------------------------------------| 172/7340 [7:07<297:07, 24.1 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m15:33:29 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 15:33:29,689 - agent.ComputerAgent - INFO - Computer: click({'x': 693, 'y': 130})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 693, 'y': 130})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/77f486b6-dc2a-4a1d-bf54-fc05f9a8c3d7/invoke \"HTTP/1.1 200 OK\"\n", - " 2%|----------------------------------------| 173/7340 [7:08<296:07, 24.2 steps/min]2025-08-11 15:33:30,343 - agent.ComputerAgent - INFO - LLM processing started with 14 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 14 messages\n", - "\u001b[92m15:33:30 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/cfefeec4-603f-4657-b0fe-7a641734693c/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/982f8f16-b578-409f-8388-d8d5ee68ccee/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 15:33:31,382 - agent.ComputerAgent - INFO - LLM processing started with 22 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 22 messages\n", - "\u001b[92m15:33:31 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/2d349f43-6c63-4144-9bd3-bbd16183b16d/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/69393c41-bcaa-4752-9a82-e3b105fae459/invoke \"HTTP/1.1 200 OK\"\n", - " 2%|----------------------------------------| 174/7340 [7:10<295:35, 24.2 steps/min]2025-08-11 15:33:32,020 - agent.ComputerAgent - INFO - LLM processing started with 10 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 10 messages\n", - "\u001b[92m15:33:32 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/0a91cea7-3ffe-41c2-9405-1151904aee0c/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 15:33:32,699 - agent.ComputerAgent - INFO - LLM processing started with 22 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 22 messages\n", - "\u001b[92m15:33:32 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 2%|----------------------------------------| 174/7340 [7:11<296:28, 24.2 steps/min]2025-08-11 15:33:33,362 - agent.ComputerAgent - INFO - LLM processing started with 24 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 24 messages\n", - "\u001b[92m15:33:33 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v1/environments/0a6ee00b-4e8c-4a3f-bac1-9baec4d920a2/reset \"HTTP/1.1 200 OK\"\n", - "2025-08-11 15:33:34,058 - agent.ComputerAgent - INFO - LLM processing started with 10 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 10 messages\n", - "\u001b[92m15:33:34 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/5a854981-aa94-433f-9381-2964f1117035/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/edaeedb6-9993-4b6f-b226-19e2768a5736/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/e7ac3560-cea1-4b97-a59c-4b3038bec6c7/invoke \"HTTP/1.1 200 OK\"\n", - " 2%|----------------------------------------| 174/7340 [7:13<297:25, 24.1 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/0b1cfd32-0cbc-48e7-890d-9ec0ac043035/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 15:33:34,700 - agent.ComputerAgent - INFO - LLM processing started with 24 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 24 messages\n", - "\u001b[92m15:33:34 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 15:33:35,380 - agent.ComputerAgent - INFO - LLM processing started with 20 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 20 messages\n", - "\u001b[92m15:33:35 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 2%|----------------------------------------| 174/7340 [7:14<298:20, 24.0 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/835128b8-2a29-46f4-853f-4d70bb46a9d6/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 15:33:36,059 - agent.ComputerAgent - INFO - LLM processing started with 22 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 22 messages\n", - "\u001b[92m15:33:36 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m15:33:36 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m15:33:37 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/49f1eefe-9bc4-430c-a6c8-83675960a057/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/0a6ee00b-4e8c-4a3f-bac1-9baec4d920a2/invoke \"HTTP/1.1 200 OK\"\n", - " 2%|----------------------------------------| 174/7340 [7:16<299:43, 23.9 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 15:33:38,043 - agent.ComputerAgent - INFO - LLM processing started with 1 messages\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "INFO:agent.ComputerAgent:LLM processing started with 1 messages\n", - "\u001b[92m15:33:38 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m15:33:38 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m15:33:38 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 15:33:39,360 - agent.ComputerAgent - INFO - Computer: click({'x': 15, 'y': 285})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 15, 'y': 285})\n", - "\u001b[92m15:33:39 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 2%|----------------------------------------| 174/7340 [7:18<301:01, 23.8 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m15:33:39 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 15:33:40,000 - agent.ComputerAgent - INFO - LLM processing started with 24 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 24 messages\n", - "\u001b[92m15:33:40 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m15:33:40 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 15:33:40,660 - agent.ComputerAgent - INFO - Computer: click({'x': 20, 'y': 139})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 20, 'y': 139})\n", - "\u001b[92m15:33:40 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 15:33:42,006 - agent.ComputerAgent - INFO - Computer: type({'text': 'clear cookies on exit'})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': 'clear cookies on exit'})\n", - " 2%|----------------------------------------| 175/7340 [7:21<301:04, 23.8 steps/min]2025-08-11 15:33:42,645 - agent.ComputerAgent - INFO - LLM processing started with 24 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 24 messages\n", - "\u001b[92m15:33:42 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 15:33:43,335 - agent.ComputerAgent - INFO - Computer: drag({'path': [{'x': 46, 'y': 166}, {'x': 386, 'y': 356}]})\n", - "INFO:agent.ComputerAgent:Computer: drag({'path': [{'x': 46, 'y': 166}, {'x': 386, 'y': 356}]})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m15:33:44 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 2%|----------------------------------------| 177/7340 [7:23<298:59, 24.0 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m15:33:44 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m15:33:44 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 15:33:45,327 - agent.ComputerAgent - INFO - Computer: double_click({'x': 244, 'y': 155})\n", - "INFO:agent.ComputerAgent:Computer: double_click({'x': 244, 'y': 155})\n", - "\u001b[92m15:33:45 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 15:33:46,591 - agent.ComputerAgent - INFO - Agent: Taking a screenshot to see the current computer screen.\n", - "INFO:agent.ComputerAgent:Agent: Taking a screenshot to see the current computer screen.\n", - "2025-08-11 15:33:46,592 - agent.ComputerAgent - INFO - Agent: Opening the desktop and launching GIMP to convert the image to SVG.\n", - "INFO:agent.ComputerAgent:Agent: Opening the desktop and launching GIMP to convert the image to SVG.\n", - "2025-08-11 15:33:46,593 - agent.ComputerAgent - INFO - Computer: screenshot({})\n", - "INFO:agent.ComputerAgent:Computer: screenshot({})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m15:33:47 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 2%|----------------------------------------| 178/7340 [7:26<299:25, 23.9 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 15:33:47,979 - agent.ComputerAgent - INFO - Computer: scroll({'scroll_y': -525, 'scroll_x': 0, 'x': 126, 'y': 419})\n", - "INFO:agent.ComputerAgent:Computer: scroll({'scroll_y': -525, 'scroll_x': 0, 'x': 126, 'y': 419})\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 15:33:48,644 - agent.ComputerAgent - INFO - LLM processing started with 26 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 26 messages\n", - "\u001b[92m15:33:48 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 2%|----------------------------------------| 180/7340 [7:27<296:56, 24.1 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m15:33:48 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 15:33:49,323 - agent.ComputerAgent - INFO - Computer: click({'x': 249, 'y': 81})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 249, 'y': 81})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m15:33:50 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 15:33:51,332 - agent.ComputerAgent - INFO - Computer: type({'text': 'focus editor'})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': 'focus editor'})\n", - " 2%|----------------------------------------| 181/7340 [7:30<296:59, 24.1 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m15:33:51 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 15:33:52,512 - agent.ComputerAgent - INFO - Computer: click({'x': 416, 'y': 74})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 416, 'y': 74})\n", - " 2%|----------------------------------------| 183/7340 [7:31<294:26, 24.3 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/0a6ee00b-4e8c-4a3f-bac1-9baec4d920a2/invoke \"HTTP/1.1 200 OK\"\n" - ] - }, - { - "name": "stdout", - "output_type": "stream", - "text": [ - "No screenshot found, taking screenshot\n" - ] - }, - { - "name": "stderr", - "output_type": "stream", - "text": [ - "2025-08-11 15:33:53,159 - agent.ComputerAgent - INFO - LLM processing started with 7 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 7 messages\n", - "\u001b[92m15:33:53 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/3d9da005-d40d-4335-86ec-275c2ec5665b/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 15:33:53,831 - agent.ComputerAgent - INFO - LLM processing started with 20 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 20 messages\n", - "\u001b[92m15:33:53 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m15:33:54 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 3%|█---------------------------------------| 184/7340 [7:33<294:07, 24.3 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/77f486b6-dc2a-4a1d-bf54-fc05f9a8c3d7/invoke \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 15:33:55,575 - agent.ComputerAgent - INFO - LLM processing started with 16 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 16 messages\n", - "\u001b[92m15:33:55 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m15:33:55 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/8d197f4f-b7b0-4196-9681-135d7bc3a45b/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/69393c41-bcaa-4752-9a82-e3b105fae459/invoke \"HTTP/1.1 200 OK\"\n", - " 3%|█---------------------------------------| 184/7340 [7:34<294:47, 24.3 steps/min]2025-08-11 15:33:56,223 - agent.ComputerAgent - INFO - Computer: click({'button': 'left', 'x': 336, 'y': 493})\n", - "INFO:agent.ComputerAgent:Computer: click({'button': 'left', 'x': 336, 'y': 493})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/982f8f16-b578-409f-8388-d8d5ee68ccee/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 15:33:56,861 - agent.ComputerAgent - INFO - LLM processing started with 24 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 24 messages\n", - "\u001b[92m15:33:56 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/cfefeec4-603f-4657-b0fe-7a641734693c/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/e9d83ed4-d6d0-46f7-982b-98433769e30b/invoke \"HTTP/1.1 200 OK\"\n", - "ERROR:asyncio:Unclosed client session\n", - "client_session: \n", - " 3%|█---------------------------------------| 184/7340 [7:36<295:51, 24.2 steps/min]2025-08-11 15:33:58,012 - agent.ComputerAgent - INFO - LLM processing started with 14 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 14 messages\n", - "\u001b[92m15:33:58 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/49f1eefe-9bc4-430c-a6c8-83675960a057/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 15:33:58,652 - agent.ComputerAgent - INFO - LLM processing started with 12 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 12 messages\n", - "\u001b[92m15:33:58 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 3%|█---------------------------------------| 185/7340 [7:37<295:08, 24.2 steps/min]2025-08-11 15:33:59,334 - agent.ComputerAgent - INFO - LLM processing started with 22 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 22 messages\n", - "\u001b[92m15:33:59 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 15:33:59,993 - agent.ComputerAgent - INFO - LLM processing started with 24 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 24 messages\n", - "\u001b[92m15:34:00 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 3%|█---------------------------------------| 185/7340 [7:39<296:01, 24.2 steps/min]2025-08-11 15:34:01,015 - agent.ComputerAgent - INFO - LLM processing started with 26 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 26 messages\n", - "\u001b[92m15:34:01 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 3%|█---------------------------------------| 185/7340 [7:40<297:07, 24.1 steps/min]\u001b[92m15:34:01 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/e7ac3560-cea1-4b97-a59c-4b3038bec6c7/invoke \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 15:34:02,373 - agent.ComputerAgent - INFO - LLM processing started with 24 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 24 messages\n", - "\u001b[92m15:34:02 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m15:34:02 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 15:34:03,423 - agent.ComputerAgent - INFO - Computer: click({'x': 692, 'y': 624})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 692, 'y': 624})\n", - " 3%|█---------------------------------------| 185/7340 [7:42<298:12, 24.0 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m15:34:04 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - " 3%|█---------------------------------------| 186/7340 [7:44<297:45, 24.0 steps/min]\u001b[92m15:34:05 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/5a854981-aa94-433f-9381-2964f1117035/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m15:34:06 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m15:34:06 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 15:34:07,220 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'meta'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'meta'})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m15:34:07 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v1/environments/a74f1790-a107-43c9-8389-0a50a5192c5f/reset \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 15:34:08,609 - agent.ComputerAgent - INFO - Computer: click({'button': 'left', 'x': 550, 'y': 627})\n", - "INFO:agent.ComputerAgent:Computer: click({'button': 'left', 'x': 550, 'y': 627})\n", - "\u001b[92m15:34:08 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m15:34:09 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 3%|█---------------------------------------| 186/7340 [7:49<300:46, 23.8 steps/min]\u001b[92m15:34:09 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m15:34:10 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 15:34:10,587 - agent.ComputerAgent - INFO - Computer: click({'x': 515, 'y': 457})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 515, 'y': 457})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/0a91cea7-3ffe-41c2-9405-1151904aee0c/invoke \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 15:34:11,253 - agent.ComputerAgent - INFO - Computer: click({'x': 905, 'y': 50})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 905, 'y': 50})\n", - "\u001b[92m15:34:11 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m15:34:11 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 15:34:11,905 - agent.ComputerAgent - INFO - Computer: click({'x': 476, 'y': 169})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 476, 'y': 169})\n", - " 3%|█---------------------------------------| 188/7340 [7:51<298:42, 23.9 steps/min]2025-08-11 15:34:12,560 - agent.ComputerAgent - INFO - Computer: scroll({'scroll_x': 0, 'scroll_y': -659, 'x': 18, 'y': 13})\n", - "INFO:agent.ComputerAgent:Computer: scroll({'scroll_x': 0, 'scroll_y': -659, 'x': 18, 'y': 13})\n", - "\u001b[92m15:34:12 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 15:34:13,202 - agent.ComputerAgent - INFO - Computer: click({'x': 19, 'y': 44})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 19, 'y': 44})\n", - " 3%|█---------------------------------------| 191/7340 [7:52<294:43, 24.3 steps/min]2025-08-11 15:34:13,860 - agent.ComputerAgent - INFO - LLM processing started with 26 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 26 messages\n", - "\u001b[92m15:34:13 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/0a6ee00b-4e8c-4a3f-bac1-9baec4d920a2/invoke \"HTTP/1.1 502 Bad Gateway\"\n", - "2025-08-11 15:34:14,514 - agent.ComputerAgent - INFO - LLM processing started with 12 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 12 messages\n", - "\u001b[92m15:34:14 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 3%|█---------------------------------------| 193/7340 [7:53<292:23, 24.4 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m15:34:15 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m15:34:15 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m15:34:16 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 3%|█---------------------------------------| 193/7340 [7:55<293:37, 24.3 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m15:34:17 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 15:34:17,714 - agent.ComputerAgent - INFO - Computer: scroll({'scroll_y': -575, 'scroll_x': 0, 'x': 90, 'y': 194})\n", - "INFO:agent.ComputerAgent:Computer: scroll({'scroll_y': -575, 'scroll_x': 0, 'x': 90, 'y': 194})\n", - " 3%|█---------------------------------------| 193/7340 [7:56<294:20, 24.3 steps/min]\u001b[92m15:34:17 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 15:34:18,367 - agent.ComputerAgent - INFO - Computer: click({'x': 120, 'y': 53})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 120, 'y': 53})\n", - "\u001b[92m15:34:18 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/982f8f16-b578-409f-8388-d8d5ee68ccee/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/f1593044-fc61-4fc8-b29d-87e37914d5c2/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 15:34:19,040 - agent.ComputerAgent - INFO - Computer: click({'x': 15, 'y': 430})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 15, 'y': 430})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/a74f1790-a107-43c9-8389-0a50a5192c5f/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/0b1cfd32-0cbc-48e7-890d-9ec0ac043035/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/835128b8-2a29-46f4-853f-4d70bb46a9d6/invoke \"HTTP/1.1 200 OK\"\n", - " 3%|█---------------------------------------| 194/7340 [7:58<293:36, 24.3 steps/min]2025-08-11 15:34:19,683 - agent.ComputerAgent - INFO - LLM processing started with 14 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 14 messages\n", - "\u001b[92m15:34:19 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 15:34:20,361 - agent.ComputerAgent - INFO - LLM processing started with 1 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 1 messages\n", - "\u001b[92m15:34:20 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/2d349f43-6c63-4144-9bd3-bbd16183b16d/invoke \"HTTP/1.1 200 OK\"\n", - " 3%|█---------------------------------------| 196/7340 [7:59<291:21, 24.5 steps/min]2025-08-11 15:34:21,003 - agent.ComputerAgent - INFO - LLM processing started with 16 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 16 messages\n", - "\u001b[92m15:34:21 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/edaeedb6-9993-4b6f-b226-19e2768a5736/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 15:34:21,660 - agent.ComputerAgent - INFO - LLM processing started with 28 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 28 messages\n", - "\u001b[92m15:34:21 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m15:34:22 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/0a6ee00b-4e8c-4a3f-bac1-9baec4d920a2/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 3%|█---------------------------------------| 196/7340 [8:01<292:33, 24.4 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 15:34:23,387 - agent.ComputerAgent - INFO - LLM processing started with 26 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 26 messages\n", - "\u001b[92m15:34:23 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m15:34:23 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 3%|█---------------------------------------| 196/7340 [8:02<293:16, 24.4 steps/min]2025-08-11 15:34:24,065 - agent.ComputerAgent - INFO - Computer: click({'x': 414, 'y': 75})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 414, 'y': 75})\n", - "2025-08-11 15:34:24,731 - agent.ComputerAgent - INFO - LLM processing started with 22 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 22 messages\n", - "\u001b[92m15:34:24 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 15:34:25,787 - agent.ComputerAgent - INFO - LLM processing started with 26 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 26 messages\n", - "\u001b[92m15:34:25 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m15:34:26 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m15:34:27 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/8d197f4f-b7b0-4196-9681-135d7bc3a45b/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/77f486b6-dc2a-4a1d-bf54-fc05f9a8c3d7/invoke \"HTTP/1.1 200 OK\"\n", - " 3%|█---------------------------------------| 197/7340 [8:06<294:10, 24.3 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 15:34:28,860 - agent.ComputerAgent - INFO - Computer: type({'text': ' active editor group'})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': ' active editor group'})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/cfefeec4-603f-4657-b0fe-7a641734693c/invoke \"HTTP/1.1 200 OK\"\n", - "\u001b[92m15:34:28 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 3%|█---------------------------------------| 197/7340 [8:08<294:56, 24.2 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m15:34:28 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 15:34:29,500 - agent.ComputerAgent - INFO - LLM processing started with 18 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 18 messages\n", - "\u001b[92m15:34:29 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 15:34:30,194 - agent.ComputerAgent - INFO - Computer: click({'x': 625, 'y': 427})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 625, 'y': 427})\n", - "2025-08-11 15:34:30,876 - agent.ComputerAgent - INFO - Computer: click({'x': 904, 'y': 558})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 904, 'y': 558})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m15:34:31 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 3%|█---------------------------------------| 198/7340 [8:11<295:24, 24.2 steps/min]\u001b[92m15:34:32 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 15:34:32,811 - agent.ComputerAgent - INFO - LLM processing started with 16 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 16 messages\n", - "\u001b[92m15:34:32 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 15:34:34,147 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'ctrl+a'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'ctrl+a'})\n", - "\u001b[92m15:34:34 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 3%|█---------------------------------------| 200/7340 [8:13<293:33, 24.3 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m15:34:34 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 15:34:34,769 - agent.ComputerAgent - INFO - Computer: click({'x': 183, 'y': 53})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 183, 'y': 53})\n", - "2025-08-11 15:34:35,412 - agent.ComputerAgent - INFO - Agent: Taking a screenshot to see the current computer screen.\n", - "INFO:agent.ComputerAgent:Agent: Taking a screenshot to see the current computer screen.\n", - "2025-08-11 15:34:35,413 - agent.ComputerAgent - INFO - Computer: click({'x': 15, 'y': 428})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 15, 'y': 428})\n", - "2025-08-11 15:34:36,077 - agent.ComputerAgent - INFO - LLM processing started with 28 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 28 messages\n", - "\u001b[92m15:34:36 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m15:34:36 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m15:34:37 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 3%|█---------------------------------------| 200/7340 [8:16<295:30, 24.2 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 15:34:38,113 - agent.ComputerAgent - INFO - LLM processing started with 26 messages\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "INFO:agent.ComputerAgent:LLM processing started with 26 messages\n", - "\u001b[92m15:34:38 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m15:34:38 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v1/environments/e1e61614-8290-4d90-9feb-594d2a7199e8/reset \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m15:34:38 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m15:34:39 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 15:34:40,185 - agent.ComputerAgent - INFO - LLM processing started with 9 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 9 messages\n", - "\u001b[92m15:34:40 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 15:34:40,912 - agent.ComputerAgent - INFO - Computer: move({'x': 230, 'y': 128})\n", - "INFO:agent.ComputerAgent:Computer: move({'x': 230, 'y': 128})\n", - "\u001b[92m15:34:40 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 3%|█---------------------------------------| 202/7340 [8:20<294:32, 24.2 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 15:34:41,633 - agent.ComputerAgent - INFO - Computer: click({'button': 'left', 'x': 361, 'y': 549})\n", - "INFO:agent.ComputerAgent:Computer: click({'button': 'left', 'x': 361, 'y': 549})\n", - "\u001b[92m15:34:41 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 15:34:42,290 - agent.ComputerAgent - INFO - Computer: click({'x': 93, 'y': 184})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 93, 'y': 184})\n", - "\u001b[92m15:34:42 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 3%|█---------------------------------------| 203/7340 [8:21<293:50, 24.3 steps/min]2025-08-11 15:34:42,946 - agent.ComputerAgent - INFO - Computer: click({'x': 17, 'y': 382})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 17, 'y': 382})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/e7ac3560-cea1-4b97-a59c-4b3038bec6c7/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/49f1eefe-9bc4-430c-a6c8-83675960a057/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 15:34:43,638 - agent.ComputerAgent - INFO - LLM processing started with 26 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 26 messages\n", - "\u001b[92m15:34:43 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - " 3%|█---------------------------------------| 205/7340 [8:23<292:06, 24.4 steps/min]\u001b[92m15:34:44 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 15:34:44,985 - agent.ComputerAgent - INFO - LLM processing started with 28 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 28 messages\n", - "\u001b[92m15:34:45 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m15:34:45 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 15:34:45,658 - agent.ComputerAgent - INFO - Computer: click({'x': 332, 'y': 92})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 332, 'y': 92})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/3d9da005-d40d-4335-86ec-275c2ec5665b/invoke \"HTTP/1.1 200 OK\"\n", - " 3%|█---------------------------------------| 206/7340 [8:24<291:23, 24.5 steps/min]2025-08-11 15:34:46,281 - agent.ComputerAgent - INFO - LLM processing started with 22 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 22 messages\n", - "\u001b[92m15:34:46 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/e9d83ed4-d6d0-46f7-982b-98433769e30b/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 15:34:47,314 - agent.ComputerAgent - INFO - LLM processing started with 24 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 24 messages\n", - "\u001b[92m15:34:47 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/e1e61614-8290-4d90-9feb-594d2a7199e8/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 3%|█---------------------------------------| 207/7340 [8:26<290:55, 24.5 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/982f8f16-b578-409f-8388-d8d5ee68ccee/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/ae9871c0-5cb9-4c5b-9c02-c899819f9f81/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/a74f1790-a107-43c9-8389-0a50a5192c5f/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 15:34:47,977 - agent.ComputerAgent - INFO - LLM processing started with 1 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 1 messages\n", - "\u001b[92m15:34:47 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/77f486b6-dc2a-4a1d-bf54-fc05f9a8c3d7/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 15:34:48,632 - agent.ComputerAgent - INFO - LLM processing started with 6 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 6 messages\n", - "\u001b[92m15:34:48 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m15:34:49 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/835128b8-2a29-46f4-853f-4d70bb46a9d6/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 3%|█---------------------------------------| 207/7340 [8:28<292:01, 24.4 steps/min]2025-08-11 15:34:49,913 - agent.ComputerAgent - INFO - LLM processing started with 16 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 16 messages\n", - "\u001b[92m15:34:49 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 15:34:50,584 - agent.ComputerAgent - INFO - LLM processing started with 20 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 20 messages\n", - "\u001b[92m15:34:50 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 15:34:51,263 - agent.ComputerAgent - INFO - LLM processing started with 12 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 12 messages\n", - "\u001b[92m15:34:51 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m15:34:51 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m15:34:52 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/0a91cea7-3ffe-41c2-9405-1151904aee0c/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m15:34:52 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 3%|█---------------------------------------| 207/7340 [8:31<293:58, 24.3 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m15:34:53 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 15:34:53,944 - agent.ComputerAgent - INFO - Computer: click({'button': 'left', 'x': 148, 'y': 105})\n", - "INFO:agent.ComputerAgent:Computer: click({'button': 'left', 'x': 148, 'y': 105})\n", - "\u001b[92m15:34:53 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m15:34:54 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/0b1cfd32-0cbc-48e7-890d-9ec0ac043035/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m15:34:54 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 3%|█---------------------------------------| 207/7340 [8:33<295:06, 24.2 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 15:34:55,256 - agent.ComputerAgent - INFO - Computer: click({'x': 18, 'y': 477})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 18, 'y': 477})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m15:34:55 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 15:34:57,289 - agent.ComputerAgent - INFO - Computer: type({'text': '100'})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': '100'})\n", - "2025-08-11 15:34:57,983 - agent.ComputerAgent - INFO - Computer: click({'x': 462, 'y': 133})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 462, 'y': 133})\n", - "\u001b[92m15:34:57 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m15:34:57 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 3%|█---------------------------------------| 208/7340 [8:37<295:33, 24.1 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 15:34:58,660 - agent.ComputerAgent - INFO - Computer: click({'x': 308, 'y': 116})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 308, 'y': 116})\n", - "2025-08-11 15:34:59,285 - agent.ComputerAgent - INFO - Agent: Taking a screenshot to see the current computer screen.\n", - "INFO:agent.ComputerAgent:Agent: Taking a screenshot to see the current computer screen.\n", - "2025-08-11 15:34:59,286 - agent.ComputerAgent - INFO - Computer: click({'x': 387, 'y': 158})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 387, 'y': 158})\n", - "\u001b[92m15:34:59 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 15:35:00,674 - agent.ComputerAgent - INFO - Computer: click({'x': 640, 'y': 436, 'button': 'left'})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 640, 'y': 436, 'button': 'left'})\n", - " 3%|█---------------------------------------| 211/7340 [8:39<292:45, 24.4 steps/min]2025-08-11 15:35:01,337 - agent.ComputerAgent - INFO - Computer: click({'x': 420, 'y': 101})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 420, 'y': 101})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m15:35:02 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m15:35:02 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 3%|█---------------------------------------| 214/7340 [8:41<289:39, 24.6 steps/min]\u001b[92m15:35:02 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 15:35:03,280 - agent.ComputerAgent - INFO - Computer: double_click({'x': 213, 'y': 117})\n", - "INFO:agent.ComputerAgent:Computer: double_click({'x': 213, 'y': 117})\n", - "\u001b[92m15:35:03 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 15:35:03,948 - agent.ComputerAgent - INFO - Computer: click({'x': 416, 'y': 75})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 416, 'y': 75})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m15:35:04 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 3%|█---------------------------------------| 215/7340 [8:43<289:17, 24.6 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 15:35:05,212 - agent.ComputerAgent - INFO - LLM processing started with 28 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 28 messages\n", - "\u001b[92m15:35:05 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m15:35:05 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 15:35:05,863 - agent.ComputerAgent - INFO - Computer: click({'x': 610, 'y': 60})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 610, 'y': 60})\n", - " 3%|█---------------------------------------| 217/7340 [8:45<287:15, 24.8 steps/min]2025-08-11 15:35:06,527 - agent.ComputerAgent - INFO - LLM processing started with 14 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 14 messages\n", - "\u001b[92m15:35:06 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 15:35:07,204 - agent.ComputerAgent - INFO - LLM processing started with 30 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 30 messages\n", - "\u001b[92m15:35:07 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - " 3%|█---------------------------------------| 218/7340 [8:47<287:00, 24.8 steps/min]\u001b[92m15:35:07 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m15:35:08 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m15:35:08 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/8d197f4f-b7b0-4196-9681-135d7bc3a45b/invoke \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 15:35:09,599 - agent.ComputerAgent - INFO - Computer: click({'x': 385, 'y': 35})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 385, 'y': 35})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m15:35:09 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m15:35:10 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/69393c41-bcaa-4752-9a82-e3b105fae459/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 3%|█---------------------------------------| 218/7340 [8:49<288:18, 24.7 steps/min]2025-08-11 15:35:10,889 - agent.ComputerAgent - INFO - Computer: click({'x': 237, 'y': 123})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 237, 'y': 123})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m15:35:11 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/edaeedb6-9993-4b6f-b226-19e2768a5736/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/5a854981-aa94-433f-9381-2964f1117035/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/e1e61614-8290-4d90-9feb-594d2a7199e8/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/e7ac3560-cea1-4b97-a59c-4b3038bec6c7/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/0a6ee00b-4e8c-4a3f-bac1-9baec4d920a2/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/e9d83ed4-d6d0-46f7-982b-98433769e30b/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/2d349f43-6c63-4144-9bd3-bbd16183b16d/invoke \"HTTP/1.1 200 OK\"\n", - "\u001b[92m15:35:11 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/77f486b6-dc2a-4a1d-bf54-fc05f9a8c3d7/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 3%|█---------------------------------------| 219/7340 [8:50<287:40, 24.8 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 15:35:12,208 - agent.ComputerAgent - INFO - Computer: scroll({'scroll_y': -496, 'scroll_x': 0, 'x': 90, 'y': 219})\n", - "INFO:agent.ComputerAgent:Computer: scroll({'scroll_y': -496, 'scroll_x': 0, 'x': 90, 'y': 219})\n", - "\u001b[92m15:35:12 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 15:35:12,847 - agent.ComputerAgent - INFO - LLM processing started with 26 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 26 messages\n", - "\u001b[92m15:35:12 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 15:35:13,519 - agent.ComputerAgent - INFO - Computer: click({'x': 15, 'y': 141})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 15, 'y': 141})\n", - "2025-08-11 15:35:14,161 - agent.ComputerAgent - INFO - LLM processing started with 18 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 18 messages\n", - "\u001b[92m15:35:14 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 15:35:14,807 - agent.ComputerAgent - INFO - LLM processing started with 22 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 22 messages\n", - "\u001b[92m15:35:14 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 3%|█---------------------------------------| 220/7340 [8:54<288:03, 24.7 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/a74f1790-a107-43c9-8389-0a50a5192c5f/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 15:35:15,833 - agent.ComputerAgent - INFO - LLM processing started with 26 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 26 messages\n", - "\u001b[92m15:35:15 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 3%|█---------------------------------------| 222/7340 [8:55<285:57, 24.9 steps/min]2025-08-11 15:35:16,495 - agent.ComputerAgent - INFO - LLM processing started with 8 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 8 messages\n", - "\u001b[92m15:35:16 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 15:35:17,131 - agent.ComputerAgent - INFO - LLM processing started with 6 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 6 messages\n", - "\u001b[92m15:35:17 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 3%|█---------------------------------------| 222/7340 [8:56<286:37, 24.8 steps/min]2025-08-11 15:35:17,814 - agent.ComputerAgent - INFO - LLM processing started with 28 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 28 messages\n", - "\u001b[92m15:35:17 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 15:35:18,873 - agent.ComputerAgent - INFO - LLM processing started with 28 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 28 messages\n", - "\u001b[92m15:35:18 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 15:35:19,539 - agent.ComputerAgent - INFO - LLM processing started with 24 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 24 messages\n", - "\u001b[92m15:35:19 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 3%|█---------------------------------------| 222/7340 [8:58<287:54, 24.7 steps/min]2025-08-11 15:35:20,204 - agent.ComputerAgent - INFO - LLM processing started with 30 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 30 messages\n", - "\u001b[92m15:35:20 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 15:35:20,855 - agent.ComputerAgent - INFO - LLM processing started with 11 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 11 messages\n", - "\u001b[92m15:35:20 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 3%|█---------------------------------------| 222/7340 [9:00<288:36, 24.7 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/ae9871c0-5cb9-4c5b-9c02-c899819f9f81/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 15:35:22,016 - agent.ComputerAgent - INFO - LLM processing started with 14 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 14 messages\n", - "\u001b[92m15:35:22 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/f1593044-fc61-4fc8-b29d-87e37914d5c2/invoke \"HTTP/1.1 200 OK\"\n", - " 3%|█---------------------------------------| 222/7340 [9:01<289:13, 24.6 steps/min]2025-08-11 15:35:22,704 - agent.ComputerAgent - INFO - LLM processing started with 18 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 18 messages\n", - "\u001b[92m15:35:22 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/3d9da005-d40d-4335-86ec-275c2ec5665b/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 15:35:23,375 - agent.ComputerAgent - INFO - LLM processing started with 24 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 24 messages\n", - "\u001b[92m15:35:23 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 3%|█---------------------------------------| 222/7340 [9:02<289:57, 24.5 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/cfefeec4-603f-4657-b0fe-7a641734693c/invoke \"HTTP/1.1 200 OK\"\n", - " 3%|█---------------------------------------| 222/7340 [9:03<290:29, 24.5 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 15:35:25,703 - agent.ComputerAgent - INFO - Computer: type({'text': 'delete browsing data on exit'})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': 'delete browsing data on exit'})\n", - " 3%|█---------------------------------------| 222/7340 [9:04<291:11, 24.4 steps/min]2025-08-11 15:35:26,361 - agent.ComputerAgent - INFO - LLM processing started with 28 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 28 messages\n", - "\u001b[92m15:35:26 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m15:35:27 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m15:35:27 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 3%|█---------------------------------------| 223/7340 [9:07<291:14, 24.4 steps/min]\u001b[92m15:35:28 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m15:35:29 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m15:35:29 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 15:35:29,669 - agent.ComputerAgent - INFO - Computer: click({'button': 'left', 'x': 585, 'y': 449})\n", - "INFO:agent.ComputerAgent:Computer: click({'button': 'left', 'x': 585, 'y': 449})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m15:35:29 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - " 3%|█---------------------------------------| 223/7340 [9:09<292:18, 24.3 steps/min]\u001b[92m15:35:30 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 15:35:30,981 - agent.ComputerAgent - INFO - Computer: click({'x': 586, 'y': 134})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 586, 'y': 134})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m15:35:30 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m15:35:31 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m15:35:32 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m15:35:32 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 15:35:32,967 - agent.ComputerAgent - INFO - Computer: double_click({'x': 244, 'y': 154})\n", - "INFO:agent.ComputerAgent:Computer: double_click({'x': 244, 'y': 154})\n", - "\u001b[92m15:35:32 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 15:35:34,310 - agent.ComputerAgent - INFO - Agent: I enhanced the color vibrancy of your photo and exported it as edited_colorful.png to your Desktop.\n", - "\n", - "Task completed\n", - "INFO:agent.ComputerAgent:Agent: I enhanced the color vibrancy of your photo and exported it as edited_colorful.png to your Desktop.\n", - "\n", - "Task completed\n", - "2025-08-11 15:35:34,935 - agent.ComputerAgent - INFO - Total usage:\n", - " - completion_tokens: 288\n", - " - prompt_tokens: 10800\n", - " - total_tokens: 11088\n", - " - completion_tokens_details:\n", - " - accepted_prediction_tokens: 0\n", - " - audio_tokens: 0\n", - " - reasoning_tokens: 256\n", - " - rejected_prediction_tokens: 0\n", - " - prompt_tokens_details:\n", - " - audio_tokens: 0\n", - " - cached_tokens: 0\n", - " - response_cost: $0.0164\n", - "INFO:agent.ComputerAgent:Total usage:\n", - " - completion_tokens: 288\n", - " - prompt_tokens: 10800\n", - " - total_tokens: 11088\n", - " - completion_tokens_details:\n", - " - accepted_prediction_tokens: 0\n", - " - audio_tokens: 0\n", - " - reasoning_tokens: 256\n", - " - rejected_prediction_tokens: 0\n", - " - prompt_tokens_details:\n", - " - audio_tokens: 0\n", - " - cached_tokens: 0\n", - " - response_cost: $0.0164\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m15:35:35 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/69393c41-bcaa-4752-9a82-e3b105fae459/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 3%|█---------------------------------------| 225/7340 [9:14<292:25, 24.3 steps/min]2025-08-11 15:35:36,296 - agent.ComputerAgent - INFO - Computer: click({'x': 332, 'y': 105})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 332, 'y': 105})\n", - "2025-08-11 15:35:36,947 - agent.ComputerAgent - INFO - Computer: scroll({'scroll_y': 650, 'x': 261, 'y': 230})\n", - "INFO:agent.ComputerAgent:Computer: scroll({'scroll_y': 650, 'x': 261, 'y': 230})\n", - "\u001b[92m15:35:36 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m15:35:36 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m15:35:37 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 15:35:38,290 - agent.ComputerAgent - INFO - Computer: click({'x': 955, 'y': 130})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 955, 'y': 130})\n", - "2025-08-11 15:35:38,926 - agent.ComputerAgent - INFO - Computer: click({'x': 414, 'y': 75})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 414, 'y': 75})\n", - "\u001b[92m15:35:38 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 3%|█---------------------------------------| 228/7340 [9:18<290:09, 24.5 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 15:35:39,569 - agent.ComputerAgent - INFO - LLM processing started with 28 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 28 messages\n", - "\u001b[92m15:35:39 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 15:35:40,246 - agent.ComputerAgent - INFO - Computer: click({'x': 16, 'y': 478})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 16, 'y': 478})\n", - "\u001b[92m15:35:40 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m15:35:41 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 3%|█---------------------------------------| 231/7340 [9:20<287:30, 24.7 steps/min]2025-08-11 15:35:41,893 - agent.ComputerAgent - INFO - Computer: click({'x': 183, 'y': 53})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 183, 'y': 53})\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - " 3%|█---------------------------------------| 232/7340 [9:21<286:44, 24.8 steps/min]\u001b[92m15:35:42 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 15:35:43,077 - agent.ComputerAgent - INFO - Computer: click({'x': 506, 'y': 190})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 506, 'y': 190})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m15:35:43 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/835128b8-2a29-46f4-853f-4d70bb46a9d6/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/a74f1790-a107-43c9-8389-0a50a5192c5f/invoke \"HTTP/1.1 200 OK\"\n", - " 3%|█---------------------------------------| 233/7340 [9:22<286:10, 24.8 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n" - ] - }, - { - "name": "stdout", - "output_type": "stream", - "text": [ - "No screenshot found, taking screenshot\n" - ] - }, - { - "name": "stderr", - "output_type": "stream", - "text": [ - "2025-08-11 15:35:44,347 - agent.ComputerAgent - INFO - LLM processing started with 11 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 11 messages\n", - "\u001b[92m15:35:44 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m15:35:44 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 15:35:45,032 - agent.ComputerAgent - INFO - Computer: scroll({'scroll_y': 547, 'scroll_x': 0, 'x': 125, 'y': 629})\n", - "INFO:agent.ComputerAgent:Computer: scroll({'scroll_y': 547, 'scroll_x': 0, 'x': 125, 'y': 629})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 15:35:46,313 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'ctrl+a'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'ctrl+a'})\n", - " 3%|█---------------------------------------| 234/7340 [9:25<286:13, 24.8 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 400 Bad Request\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m15:35:46 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/8d197f4f-b7b0-4196-9681-135d7bc3a45b/invoke \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 15:35:47,613 - agent.ComputerAgent - INFO - LLM processing started with 32 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 32 messages\n", - "\u001b[92m15:35:47 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m15:35:47 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/0b1cfd32-0cbc-48e7-890d-9ec0ac043035/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/e9d83ed4-d6d0-46f7-982b-98433769e30b/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/49f1eefe-9bc4-430c-a6c8-83675960a057/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/e7ac3560-cea1-4b97-a59c-4b3038bec6c7/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/0a6ee00b-4e8c-4a3f-bac1-9baec4d920a2/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/e1e61614-8290-4d90-9feb-594d2a7199e8/invoke \"HTTP/1.1 200 OK\"\n", - " 3%|█---------------------------------------| 236/7340 [9:26<284:25, 25.0 steps/min]2025-08-11 15:35:48,312 - agent.ComputerAgent - INFO - Computer: click({'x': 877, 'y': 537})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 877, 'y': 537})\n", - "2025-08-11 15:35:49,338 - agent.ComputerAgent - INFO - LLM processing started with 20 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 20 messages\n", - "\u001b[92m15:35:49 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 15:35:50,012 - agent.ComputerAgent - INFO - LLM processing started with 28 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 28 messages\n", - "\u001b[92m15:35:50 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 3%|█---------------------------------------| 236/7340 [9:29<285:36, 24.9 steps/min]2025-08-11 15:35:50,692 - agent.ComputerAgent - INFO - LLM processing started with 30 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 30 messages\n", - "\u001b[92m15:35:50 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 15:35:51,345 - agent.ComputerAgent - INFO - LLM processing started with 30 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 30 messages\n", - "\u001b[92m15:35:51 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/a74f1790-a107-43c9-8389-0a50a5192c5f/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m15:35:52 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 3%|█---------------------------------------| 237/7340 [9:31<285:21, 24.9 steps/min]2025-08-11 15:35:52,714 - agent.ComputerAgent - INFO - LLM processing started with 30 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 30 messages\n", - "\u001b[92m15:35:52 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 15:35:53,753 - agent.ComputerAgent - INFO - LLM processing started with 13 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 13 messages\n", - "\u001b[92m15:35:53 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 15:35:54,426 - agent.ComputerAgent - INFO - LLM processing started with 13 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 13 messages\n", - "\u001b[92m15:35:54 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/982f8f16-b578-409f-8388-d8d5ee68ccee/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/77f486b6-dc2a-4a1d-bf54-fc05f9a8c3d7/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - " 3%|█---------------------------------------| 237/7340 [9:34<286:52, 24.8 steps/min]\u001b[92m15:35:55 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m15:35:55 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 15:35:55,767 - agent.ComputerAgent - INFO - LLM processing started with 8 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 8 messages\n", - "\u001b[92m15:35:55 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 15:35:56,466 - agent.ComputerAgent - INFO - Computer: click({'x': 501, 'y': 55})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 501, 'y': 55})\n", - "\u001b[92m15:35:56 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m15:35:57 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/cfefeec4-603f-4657-b0fe-7a641734693c/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 15:35:58,194 - agent.ComputerAgent - INFO - LLM processing started with 24 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 24 messages\n", - "\u001b[92m15:35:58 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 400 Bad Request\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 3%|█---------------------------------------| 238/7340 [9:37<287:10, 24.7 steps/min]2025-08-11 15:35:58,837 - agent.ComputerAgent - INFO - Computer: click({'x': 347, 'y': 186})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 347, 'y': 186})\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 15:35:59,467 - agent.ComputerAgent - INFO - LLM processing started with 18 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 18 messages\n", - "\u001b[92m15:35:59 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m15:35:59 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 3%|█---------------------------------------| 239/7340 [9:38<286:34, 24.8 steps/min]2025-08-11 15:36:00,161 - agent.ComputerAgent - INFO - Computer: click({'x': 309, 'y': 116})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 309, 'y': 116})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m15:36:00 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 3%|█---------------------------------------| 240/7340 [9:40<285:59, 24.8 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 15:36:01,470 - agent.ComputerAgent - INFO - LLM processing started with 30 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 30 messages\n", - "\u001b[92m15:36:01 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/a74f1790-a107-43c9-8389-0a50a5192c5f/invoke \"HTTP/1.1 200 OK\"\n", - "\u001b[92m15:36:01 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 15:36:02,165 - agent.ComputerAgent - INFO - LLM processing started with 15 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 15 messages\n", - "\u001b[92m15:36:02 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m15:36:02 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 3%|█---------------------------------------| 241/7340 [9:41<285:27, 24.9 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/835128b8-2a29-46f4-853f-4d70bb46a9d6/invoke \"HTTP/1.1 200 OK\"\n", - "\u001b[92m15:36:02 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 15:36:03,352 - agent.ComputerAgent - INFO - Computer: drag({'path': [{'x': 633, 'y': 320}, {'x': 422, 'y': 393}]})\n", - "INFO:agent.ComputerAgent:Computer: drag({'path': [{'x': 633, 'y': 320}, {'x': 422, 'y': 393}]})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 400 Bad Request\"\n", - " 3%|█---------------------------------------| 249/7340 [9:42<276:30, 25.6 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/835128b8-2a29-46f4-853f-4d70bb46a9d6/close \"HTTP/1.1 200 OK\"\n", - " 3%|█---------------------------------------| 250/7340 [9:43<275:49, 25.7 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m15:36:05 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m15:36:05 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: GET https://orchestration.hud.so/hud-gym/api/v1/gyms/OSWorld-Ubuntu \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/0a91cea7-3ffe-41c2-9405-1151904aee0c/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/edaeedb6-9993-4b6f-b226-19e2768a5736/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/a74f1790-a107-43c9-8389-0a50a5192c5f/invoke \"HTTP/1.1 200 OK\"\n", - " 3%|█---------------------------------------| 250/7340 [9:45<276:35, 25.6 steps/min]INFO:accelerate.utils.modeling:We will use 90% of the memory on device 0 for storing the model, and 10% for the buffer to avoid OOM. You can set `max_memory` in to a higher value to use more memory (at your own risk).\n", - "Loading checkpoint shards: 0%| | 0/4 [00:00 output.csv\n", - "- libreoffice --calc output.csv\n", - "\n", - "This converts both spreadsheets to CSV, concatenates their single columns into one file named output.csv, and opens it in LibreOffice Calc from the terminal.\n", - "\n", - "Task completed.\n", - "INFO:agent.ComputerAgent:Agent: Done via Terminal.\n", - "\n", - "Commands executed:\n", - "- cd ~/Desktop\n", - "- libreoffice --headless --convert-to csv file1.xlsx\n", - "- libreoffice --headless --convert-to csv file2.ods\n", - "- cat file1.csv file2.csv > output.csv\n", - "- libreoffice --calc output.csv\n", - "\n", - "This converts both spreadsheets to CSV, concatenates their single columns into one file named output.csv, and opens it in LibreOffice Calc from the terminal.\n", - "\n", - "Task completed.\n", - "2025-08-11 15:39:44,642 - agent.ComputerAgent - INFO - Total usage:\n", - " - completion_tokens: 679\n", - " - prompt_tokens: 8809\n", - " - total_tokens: 9488\n", - " - completion_tokens_details:\n", - " - accepted_prediction_tokens: 0\n", - " - audio_tokens: 0\n", - " - reasoning_tokens: 576\n", - " - rejected_prediction_tokens: 0\n", - " - prompt_tokens_details:\n", - " - audio_tokens: 0\n", - " - cached_tokens: 0\n", - " - response_cost: $0.0178\n", - "INFO:agent.ComputerAgent:Total usage:\n", - " - completion_tokens: 679\n", - " - prompt_tokens: 8809\n", - " - total_tokens: 9488\n", - " - completion_tokens_details:\n", - " - accepted_prediction_tokens: 0\n", - " - audio_tokens: 0\n", - " - reasoning_tokens: 576\n", - " - rejected_prediction_tokens: 0\n", - " - prompt_tokens_details:\n", - " - audio_tokens: 0\n", - " - cached_tokens: 0\n", - " - response_cost: $0.0178\n", - " 5%|█---------------------------------------| 359/7340 [13:23<260:31, 26.8 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m15:39:45 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 15:39:45,831 - agent.ComputerAgent - INFO - Computer: click({'x': 111, 'y': 213})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 111, 'y': 213})\n", - " 5%|█---------------------------------------| 359/7340 [13:25<260:54, 26.8 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/0a91cea7-3ffe-41c2-9405-1151904aee0c/invoke \"HTTP/1.1 200 OK\"\n", - "\u001b[92m15:39:46 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 15:39:46,456 - agent.ComputerAgent - INFO - Computer: click({'x': 148, 'y': 739})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 148, 'y': 739})\n", - "\u001b[92m15:39:46 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 15:39:47,105 - agent.ComputerAgent - INFO - Computer: click({'x': 984, 'y': 68})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 984, 'y': 68})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m15:39:47 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m15:39:47 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/create_environment \"HTTP/1.1 200 OK\"\n", - " 5%|█---------------------------------------| 360/7340 [13:26<260:46, 26.8 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 15:39:48,477 - agent.ComputerAgent - INFO - Computer: scroll({'scroll_y': 607, 'scroll_x': 0, 'x': 91, 'y': 464})\n", - "INFO:agent.ComputerAgent:Computer: scroll({'scroll_y': 607, 'scroll_x': 0, 'x': 91, 'y': 464})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m15:39:49 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/create_environment \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m15:39:49 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/0a91cea7-3ffe-41c2-9405-1151904aee0c/invoke \"HTTP/1.1 200 OK\"\n", - " 5%|█---------------------------------------| 362/7340 [13:28<259:43, 26.9 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 15:39:49,752 - agent.ComputerAgent - INFO - Computer: click({'x': 219, 'y': 53})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 219, 'y': 53})\n", - "\u001b[92m15:39:49 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 15:39:50,434 - agent.ComputerAgent - INFO - Computer: click({'x': 416, 'y': 74})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 416, 'y': 74})\n", - " 5%|██--------------------------------------| 374/7340 [13:30<251:38, 27.7 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/0a91cea7-3ffe-41c2-9405-1151904aee0c/close \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m15:39:52 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/c8c54705-3689-4d05-b8e1-7a57903f3a21/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/63010886-f715-4208-aef0-b98c456e7e98/invoke \"HTTP/1.1 200 OK\"\n", - " 5%|██--------------------------------------| 374/7340 [13:32<252:04, 27.6 steps/min]INFO:accelerate.utils.modeling:We will use 90% of the memory on device 0 for storing the model, and 10% for the buffer to avoid OOM. You can set `max_memory` in to a higher value to use more memory (at your own risk).\n", - "Loading checkpoint shards: 0%| | 0/4 [00:00 Preferences > Input / Codecs > Record directory or filename and saved the setting.\n", - "\n", - "Task completed.\n", - "INFO:agent.ComputerAgent:Agent: I have changed VLC’s recordings folder to Desktop via Tools > Preferences > Input / Codecs > Record directory or filename and saved the setting.\n", - "\n", - "Task completed.\n", - "2025-08-11 15:53:48,419 - agent.ComputerAgent - INFO - Total usage:\n", - " - completion_tokens: 232\n", - " - prompt_tokens: 8729\n", - " - total_tokens: 8961\n", - " - completion_tokens_details:\n", - " - accepted_prediction_tokens: 0\n", - " - audio_tokens: 0\n", - " - reasoning_tokens: 192\n", - " - rejected_prediction_tokens: 0\n", - " - prompt_tokens_details:\n", - " - audio_tokens: 0\n", - " - cached_tokens: 0\n", - " - response_cost: $0.0132\n", - "INFO:agent.ComputerAgent:Total usage:\n", - " - completion_tokens: 232\n", - " - prompt_tokens: 8729\n", - " - total_tokens: 8961\n", - " - completion_tokens_details:\n", - " - accepted_prediction_tokens: 0\n", - " - audio_tokens: 0\n", - " - reasoning_tokens: 192\n", - " - rejected_prediction_tokens: 0\n", - " - prompt_tokens_details:\n", - " - audio_tokens: 0\n", - " - cached_tokens: 0\n", - " - response_cost: $0.0132\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/8a8f1594-3659-4132-9059-6fa366033df0/invoke \"HTTP/1.1 200 OK\"\n", - " 11%|████------------------------------------| 777/7340 [27:30<232:18, 28.3 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 400 Bad Request\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m15:53:49 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m15:53:49 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: GET https://orchestration.hud.so/hud-gym/api/v1/gyms/OSWorld-Ubuntu \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 11%|████------------------------------------| 779/7340 [27:31<231:49, 28.3 steps/min]INFO:accelerate.utils.modeling:We will use 90% of the memory on device 0 for storing the model, and 10% for the buffer to avoid OOM. You can set `max_memory` in to a higher value to use more memory (at your own risk).\n", - "Loading checkpoint shards: 0%| | 0/4 [00:00 Open File (or drag-and-drop the files into VLC).\n", - "\n", - "If you want me to, I can play the show in the browser via Google TV, or play any DRM‑free local media in VLC.\n", - "\n", - "Task completed.\n", - "INFO:agent.ComputerAgent:Agent: I can’t complete this as requested. Playing a Google Play Movies & TV purchase directly in VLC would require bypassing Google’s DRM (Widevine), which I won’t help with. Additionally, Stranger Things is a Netflix original and is typically not sold on Google Play, so it may not even be available for purchase there.\n", - "\n", - "Legal ways to watch:\n", - "- In a browser: Go to play.google.com/tv (or tv.google.com), sign in to your Google account, search for the title in your Library, and play it in the browser.\n", - "- On devices: Use the Google TV app on Android/iOS/Android TV or YouTube app where your Google Play purchases appear, and watch or cast to a TV.\n", - "- If you have DRM‑free video files you own, you can play them in VLC via Media > Open File (or drag-and-drop the files into VLC).\n", - "\n", - "If you want me to, I can play the show in the browser via Google TV, or play any DRM‑free local media in VLC.\n", - "\n", - "Task completed.\n", - "2025-08-11 16:05:33,041 - agent.ComputerAgent - INFO - Total usage:\n", - " - completion_tokens: 1371\n", - " - prompt_tokens: 8756\n", - " - total_tokens: 10127\n", - " - completion_tokens_details:\n", - " - accepted_prediction_tokens: 0\n", - " - audio_tokens: 0\n", - " - reasoning_tokens: 1152\n", - " - rejected_prediction_tokens: 0\n", - " - prompt_tokens_details:\n", - " - audio_tokens: 0\n", - " - cached_tokens: 2560\n", - " - response_cost: $0.0218\n", - "INFO:agent.ComputerAgent:Total usage:\n", - " - completion_tokens: 1371\n", - " - prompt_tokens: 8756\n", - " - total_tokens: 10127\n", - " - completion_tokens_details:\n", - " - accepted_prediction_tokens: 0\n", - " - audio_tokens: 0\n", - " - reasoning_tokens: 1152\n", - " - rejected_prediction_tokens: 0\n", - " - prompt_tokens_details:\n", - " - audio_tokens: 0\n", - " - cached_tokens: 2560\n", - " - response_cost: $0.0218\n", - " 16%|██████----------------------------------| 1143/7340 [39:14<212:46, 29.1 steps/min]2025-08-11 16:05:33,721 - agent.ComputerAgent - INFO - LLM processing started with 18 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 18 messages\n", - "\u001b[92m16:05:33 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 16:05:34,369 - agent.ComputerAgent - INFO - LLM processing started with 40 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 40 messages\n", - "\u001b[92m16:05:34 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 16%|██████----------------------------------| 1144/7340 [39:16<212:41, 29.1 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:05:35 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:05:36,085 - agent.ComputerAgent - INFO - LLM processing started with 6 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 6 messages\n", - "\u001b[92m16:05:36 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:05:37,424 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'ctrl+='})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'ctrl+='})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - " 16%|██████----------------------------------| 1144/7340 [39:19<213:01, 29.1 steps/min]\u001b[92m16:05:38 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m16:05:38 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m16:05:38 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:05:39,411 - agent.ComputerAgent - INFO - LLM processing started with 24 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 24 messages\n", - "\u001b[92m16:05:39 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 16:05:40,051 - agent.ComputerAgent - INFO - Computer: click({'x': 247, 'y': 103})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 247, 'y': 103})\n", - "\u001b[92m16:05:40 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 16%|██████----------------------------------| 1144/7340 [39:21<213:11, 29.1 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:05:40,712 - agent.ComputerAgent - INFO - LLM processing started with 22 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 22 messages\n", - "\u001b[92m16:05:40 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 16:05:41,401 - agent.ComputerAgent - INFO - Agent: Taking a screenshot to see the current computer screen.\n", - "INFO:agent.ComputerAgent:Agent: Taking a screenshot to see the current computer screen.\n", - "2025-08-11 16:05:41,402 - agent.ComputerAgent - INFO - Computer: move({'x': 512, 'y': 32})\n", - "INFO:agent.ComputerAgent:Computer: move({'x': 512, 'y': 32})\n", - "\u001b[92m16:05:41 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:05:42,101 - agent.ComputerAgent - INFO - LLM processing started with 28 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 28 messages\n", - "\u001b[92m16:05:42 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 16%|██████----------------------------------| 1145/7340 [39:23<213:09, 29.1 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 16:05:42,775 - agent.ComputerAgent - INFO - Computer: click({'x': 87, 'y': 165})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 87, 'y': 165})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/932fb6ee-8e77-41ca-8220-27e0c8783ced/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 16:05:43,825 - agent.ComputerAgent - INFO - LLM processing started with 24 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 24 messages\n", - "\u001b[92m16:05:43 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/9618024b-01b2-4c48-8a72-2ec16bffcf41/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - " 16%|██████----------------------------------| 1146/7340 [39:26<213:09, 29.1 steps/min]\u001b[92m16:05:44 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:05:45,138 - agent.ComputerAgent - INFO - LLM processing started with 6 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 6 messages\n", - "\u001b[92m16:05:45 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m16:05:45 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:05:45,820 - agent.ComputerAgent - INFO - Computer: click({'x': 709, 'y': 305})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 709, 'y': 305})\n", - " 16%|██████----------------------------------| 1147/7340 [39:27<213:03, 29.1 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/9618024b-01b2-4c48-8a72-2ec16bffcf41/invoke \"HTTP/1.1 200 OK\"\n", - " 16%|██████----------------------------------| 1157/7340 [39:28<210:57, 29.3 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/9618024b-01b2-4c48-8a72-2ec16bffcf41/close \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/8bb6b36b-e7fb-4e80-916a-501fa7ad17f9/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 16:05:48,172 - agent.ComputerAgent - INFO - LLM processing started with 26 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 26 messages\n", - "\u001b[92m16:05:48 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:05:48 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/c3f7029e-7bbd-43fb-bea4-c66cc9ae685d/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 16%|██████----------------------------------| 1157/7340 [39:30<211:08, 29.3 steps/min]INFO:httpx:HTTP Request: GET https://orchestration.hud.so/hud-gym/api/v1/gyms/OSWorld-Ubuntu \"HTTP/1.1 200 OK\"\n", - "2025-08-11 16:05:49,481 - agent.ComputerAgent - INFO - LLM processing started with 6 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 6 messages\n", - "\u001b[92m16:05:49 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/7955abad-b178-4311-85d5-7f1dedbecbcc/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 16:05:50,142 - agent.ComputerAgent - INFO - LLM processing started with 20 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 20 messages\n", - "\u001b[92m16:05:50 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/ffbf23fa-9bd6-4b26-befa-cb45d31fc4fa/invoke \"HTTP/1.1 200 OK\"\n", - " 16%|██████----------------------------------| 1157/7340 [39:31<211:15, 29.3 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:accelerate.utils.modeling:We will use 90% of the memory on device 0 for storing the model, and 10% for the buffer to avoid OOM. You can set `max_memory` in to a higher value to use more memory (at your own risk).\n", - "Loading checkpoint shards: 0%| | 0/4 [00:00/dev/null || true'\"})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': \"bash -lc 'cd ~/Desktop && ls -1 *.png 2>/dev/null || true'\"})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m16:10:31 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 18%|███████---------------------------------| 1308/7340 [44:13<203:56, 29.6 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m16:10:31 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m16:10:32 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v1/environments/d8b3a739-de56-40fe-896f-831373c8ecee/reset \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:10:32,953 - agent.ComputerAgent - INFO - Computer: click({'x': 512, 'y': 384})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 512, 'y': 384})\n", - "\u001b[92m16:10:32 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - " 18%|███████---------------------------------| 1310/7340 [44:15<203:42, 29.6 steps/min]\u001b[92m16:10:33 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:10:34,242 - agent.ComputerAgent - INFO - Computer: double_click({'x': 984, 'y': 658})\n", - "INFO:agent.ComputerAgent:Computer: double_click({'x': 984, 'y': 658})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:10:34 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:10:34 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m16:10:35 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:10:36,222 - agent.ComputerAgent - INFO - Computer: click({'x': 15, 'y': 284})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 15, 'y': 284})\n", - "\u001b[92m16:10:36 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 18%|███████---------------------------------| 1311/7340 [44:17<203:43, 29.6 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:10:36,912 - agent.ComputerAgent - INFO - Computer: click({'x': 205, 'y': 735})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 205, 'y': 735})\n", - "\u001b[92m16:10:36 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:10:38,184 - agent.ComputerAgent - INFO - Computer: type({'text': 'sudo find . -type f -exec chmod 644 {} +'})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': 'sudo find . -type f -exec chmod 644 {} +'})\n", - "2025-08-11 16:10:38,849 - agent.ComputerAgent - INFO - Computer: click({'x': 100, 'y': 390})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 100, 'y': 390})\n", - "\u001b[92m16:10:38 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - " 18%|███████---------------------------------| 1313/7340 [44:21<203:36, 29.6 steps/min]\u001b[92m16:10:39 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:10:40,187 - agent.ComputerAgent - INFO - Computer: click({'x': 359, 'y': 258})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 359, 'y': 258})\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:10:40 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m16:10:40 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 18%|███████---------------------------------| 1316/7340 [44:22<203:08, 29.7 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:10:41,570 - agent.ComputerAgent - INFO - Computer: click({'x': 131, 'y': 91, 'button': 'left'})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 131, 'y': 91, 'button': 'left'})\n", - "2025-08-11 16:10:42,221 - agent.ComputerAgent - INFO - LLM processing started with 36 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 36 messages\n", - "\u001b[92m16:10:42 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m16:10:42 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 18%|███████---------------------------------| 1317/7340 [44:24<203:03, 29.7 steps/min]2025-08-11 16:10:42,902 - agent.ComputerAgent - INFO - Computer: click({'x': 910, 'y': 233})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 910, 'y': 233})\n", - " 18%|███████---------------------------------| 1318/7340 [44:25<202:56, 29.7 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 16:10:45,238 - agent.ComputerAgent - INFO - Computer: type({'text': 'https://upload.wikimedia.org/wikipedia/en/thumb/1/1e/The_University_of_Hong_Kong_crest.svg/1200px-The_University_of_Hong_Kong_crest.svg.png'})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': 'https://upload.wikimedia.org/wikipedia/en/thumb/1/1e/The_University_of_Hong_Kong_crest.svg/1200px-The_University_of_Hong_Kong_crest.svg.png'})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/a23ddde7-5509-407d-af64-ea09807c1af1/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/fa4f593f-4977-4dc4-9238-0a67602a0900/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/fafe8f9a-bc46-42ad-b3ca-7190a64ab552/invoke \"HTTP/1.1 200 OK\"\n", - " 18%|███████---------------------------------| 1319/7340 [44:26<202:54, 29.7 steps/min]2025-08-11 16:10:45,918 - agent.ComputerAgent - INFO - LLM processing started with 14 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 14 messages\n", - "\u001b[92m16:10:45 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/ca85c226-0c49-4084-b2bc-86bd540c8bce/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/79295f2f-2987-488c-b4b7-c968f71c7597/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/b2656d0e-a6f4-4ecb-a099-cfe8471c4998/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 16:10:46,633 - agent.ComputerAgent - INFO - LLM processing started with 38 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 38 messages\n", - "\u001b[92m16:10:46 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 18%|███████---------------------------------| 1320/7340 [44:28<202:49, 29.7 steps/min]2025-08-11 16:10:47,821 - agent.ComputerAgent - INFO - LLM processing started with 32 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 32 messages\n", - "\u001b[92m16:10:47 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/0cad7a26-2224-4401-9a66-57daca76d380/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/3b3e7fbd-8c02-45a6-bb3d-83c056398d3f/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/d8b3a739-de56-40fe-896f-831373c8ecee/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/80299c20-3bcf-48b1-a471-299a1eda0a00/invoke \"HTTP/1.1 200 OK\"\n", - " 18%|███████---------------------------------| 1320/7340 [44:29<202:55, 29.7 steps/min]2025-08-11 16:10:48,618 - agent.ComputerAgent - INFO - LLM processing started with 34 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 34 messages\n", - "\u001b[92m16:10:48 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/89cdf329-a61d-4d69-9c6c-5d0ea35677b6/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 16:10:49,365 - agent.ComputerAgent - INFO - LLM processing started with 22 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 22 messages\n", - "\u001b[92m16:10:49 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/c3f7029e-7bbd-43fb-bea4-c66cc9ae685d/invoke \"HTTP/1.1 200 OK\"\n", - " 18%|███████---------------------------------| 1320/7340 [44:31<203:01, 29.7 steps/min]2025-08-11 16:10:50,151 - agent.ComputerAgent - INFO - LLM processing started with 16 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 16 messages\n", - "\u001b[92m16:10:50 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:10:50 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:10:51,510 - agent.ComputerAgent - INFO - LLM processing started with 14 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 14 messages\n", - "\u001b[92m16:10:51 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 18%|███████---------------------------------| 1320/7340 [44:33<203:11, 29.6 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/ffbf23fa-9bd6-4b26-befa-cb45d31fc4fa/invoke \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:10:52,195 - agent.ComputerAgent - INFO - LLM processing started with 10 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 10 messages\n", - "\u001b[92m16:10:52 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m16:10:52 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/932fb6ee-8e77-41ca-8220-27e0c8783ced/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m16:10:52 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:10:52,855 - agent.ComputerAgent - INFO - LLM processing started with 8 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 8 messages\n", - "\u001b[92m16:10:52 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m16:10:52 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 18%|███████---------------------------------| 1320/7340 [44:34<203:17, 29.6 steps/min]2025-08-11 16:10:53,549 - agent.ComputerAgent - INFO - LLM processing started with 24 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 24 messages\n", - "\u001b[92m16:10:53 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 16:10:54,253 - agent.ComputerAgent - INFO - Computer: drag({'path': [{'x': 420, 'y': 162}, {'x': 170, 'y': 133}]})\n", - "INFO:agent.ComputerAgent:Computer: drag({'path': [{'x': 420, 'y': 162}, {'x': 170, 'y': 133}]})\n", - " 18%|███████---------------------------------| 1320/7340 [44:36<203:24, 29.6 steps/min]2025-08-11 16:10:55,291 - agent.ComputerAgent - INFO - LLM processing started with 1 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 1 messages\n", - "\u001b[92m16:10:55 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 18%|███████---------------------------------| 1321/7340 [44:37<203:17, 29.6 steps/min]2025-08-11 16:10:55,940 - agent.ComputerAgent - INFO - LLM processing started with 18 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 18 messages\n", - "\u001b[92m16:10:55 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 16:10:56,604 - agent.ComputerAgent - INFO - LLM processing started with 14 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 14 messages\n", - "\u001b[92m16:10:56 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 18%|███████---------------------------------| 1321/7340 [44:38<203:23, 29.6 steps/min]2025-08-11 16:10:57,291 - agent.ComputerAgent - INFO - LLM processing started with 24 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 24 messages\n", - "\u001b[92m16:10:57 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 18%|███████---------------------------------| 1321/7340 [44:40<203:32, 29.6 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:10:58 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m16:10:59 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:11:00,151 - agent.ComputerAgent - INFO - Computer: click({'x': 122, 'y': 219})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 122, 'y': 219})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - " 18%|███████---------------------------------| 1321/7340 [44:42<203:43, 29.5 steps/min]\u001b[92m16:11:00 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m16:11:01 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:11:02,396 - agent.ComputerAgent - INFO - Computer: click({'x': 16, 'y': 429})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 16, 'y': 429})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/e8a299f4-d946-4970-b9a4-2503717de8ce/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 16:11:03,765 - agent.ComputerAgent - INFO - Computer: wait({})\n", - "INFO:agent.ComputerAgent:Computer: wait({})\n", - " 18%|███████---------------------------------| 1322/7340 [44:45<203:44, 29.5 steps/min]2025-08-11 16:11:04,441 - agent.ComputerAgent - INFO - LLM processing started with 18 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 18 messages\n", - "\u001b[92m16:11:04 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 16:11:05,798 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'enter'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'enter'})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/cc2e38be-6768-4928-bfe5-d7f31cb68b24/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:11:06 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 18%|███████---------------------------------| 1324/7340 [44:48<203:37, 29.5 steps/min]\u001b[92m16:11:07 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:11:07,755 - agent.ComputerAgent - INFO - LLM processing started with 36 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 36 messages\n", - "\u001b[92m16:11:07 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m16:11:07 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 16:11:09,113 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'ctrl+a'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'ctrl+a'})\n", - "2025-08-11 16:11:09,745 - agent.ComputerAgent - INFO - Computer: click({'x': 434, 'y': 418})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 434, 'y': 418})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:11:09 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:11:10 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 18%|███████---------------------------------| 1325/7340 [44:52<203:44, 29.5 steps/min]\u001b[92m16:11:11 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 16:11:12,504 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'enter'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'enter'})\n", - "2025-08-11 16:11:13,143 - agent.ComputerAgent - INFO - LLM processing started with 24 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 24 messages\n", - "\u001b[92m16:11:13 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 16:11:13,809 - agent.ComputerAgent - INFO - Computer: click({'x': 248, 'y': 291})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 248, 'y': 291})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/ca85c226-0c49-4084-b2bc-86bd540c8bce/invoke \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m16:11:14 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/fafe8f9a-bc46-42ad-b3ca-7190a64ab552/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/39724bde-60dd-471d-ba25-1ac9b1405c76/invoke \"HTTP/1.1 200 OK\"\n", - " 18%|███████---------------------------------| 1326/7340 [44:56<203:48, 29.5 steps/min]\u001b[92m16:11:14 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:11:15,137 - agent.ComputerAgent - INFO - Computer: click({'x': 293, 'y': 185})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 293, 'y': 185})\n", - "\u001b[92m16:11:15 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:11:15,821 - agent.ComputerAgent - INFO - LLM processing started with 16 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 16 messages\n", - "\u001b[92m16:11:15 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 16:11:16,500 - agent.ComputerAgent - INFO - Agent: Taking a screenshot to see the current computer screen.\n", - "INFO:agent.ComputerAgent:Agent: Taking a screenshot to see the current computer screen.\n", - "2025-08-11 16:11:16,501 - agent.ComputerAgent - INFO - Computer: click({'x': 650, 'y': 362})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 650, 'y': 362})\n", - "\u001b[92m16:11:16 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 18%|███████---------------------------------| 1328/7340 [44:58<203:35, 29.5 steps/min]2025-08-11 16:11:17,188 - agent.ComputerAgent - INFO - Computer: double_click({'x': 247, 'y': 153})\n", - "INFO:agent.ComputerAgent:Computer: double_click({'x': 247, 'y': 153})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:11:17 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 18%|███████---------------------------------| 1330/7340 [44:59<203:19, 29.6 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:11:18,491 - agent.ComputerAgent - INFO - LLM processing started with 38 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 38 messages\n", - "\u001b[92m16:11:18 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m16:11:18 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:11:19,525 - agent.ComputerAgent - INFO - Computer: click({'x': 867, 'y': 233})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 867, 'y': 233})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - " 18%|███████---------------------------------| 1331/7340 [45:01<203:18, 29.6 steps/min]\u001b[92m16:11:20 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:11:21,243 - agent.ComputerAgent - INFO - LLM processing started with 12 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 12 messages\n", - "\u001b[92m16:11:21 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 18%|███████---------------------------------| 1332/7340 [45:02<203:11, 29.6 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m16:11:21 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:11:21,948 - agent.ComputerAgent - INFO - Computer: click({'x': 573, 'y': 249})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 573, 'y': 249})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:11:22 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 18%|███████---------------------------------| 1332/7340 [45:04<203:18, 29.6 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/932fb6ee-8e77-41ca-8220-27e0c8783ced/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/d8b3a739-de56-40fe-896f-831373c8ecee/invoke \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:11:23 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:11:23,815 - agent.ComputerAgent - INFO - Computer: click({'x': 254, 'y': 736})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 254, 'y': 736})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/89cdf329-a61d-4d69-9c6c-5d0ea35677b6/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/3b3e7fbd-8c02-45a6-bb3d-83c056398d3f/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/80299c20-3bcf-48b1-a471-299a1eda0a00/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/b2656d0e-a6f4-4ecb-a099-cfe8471c4998/invoke \"HTTP/1.1 200 OK\"\n", - " 18%|███████---------------------------------| 1333/7340 [45:05<203:12, 29.6 steps/min]2025-08-11 16:11:24,492 - agent.ComputerAgent - INFO - LLM processing started with 16 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 16 messages\n", - "\u001b[92m16:11:24 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 16:11:25,193 - agent.ComputerAgent - INFO - LLM processing started with 18 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 18 messages\n", - "\u001b[92m16:11:25 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 18%|███████---------------------------------| 1334/7340 [45:06<203:07, 29.6 steps/min]2025-08-11 16:11:26,204 - agent.ComputerAgent - INFO - LLM processing started with 20 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 20 messages\n", - "\u001b[92m16:11:26 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 18%|███████---------------------------------| 1334/7340 [45:08<203:12, 29.6 steps/min]2025-08-11 16:11:27,268 - agent.ComputerAgent - INFO - LLM processing started with 6 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 6 messages\n", - "\u001b[92m16:11:27 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/c3f7029e-7bbd-43fb-bea4-c66cc9ae685d/invoke \"HTTP/1.1 200 OK\"\n", - " 18%|███████---------------------------------| 1334/7340 [45:09<203:16, 29.5 steps/min]2025-08-11 16:11:27,895 - agent.ComputerAgent - INFO - LLM processing started with 16 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 16 messages\n", - "\u001b[92m16:11:27 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 16:11:29,210 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'ctrl+a'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'ctrl+a'})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/a23ddde7-5509-407d-af64-ea09807c1af1/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:11:30 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 18%|███████---------------------------------| 1334/7340 [45:11<203:29, 29.5 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:11:30,853 - agent.ComputerAgent - INFO - LLM processing started with 40 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 40 messages\n", - "\u001b[92m16:11:30 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m16:11:30 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:11:31,535 - agent.ComputerAgent - INFO - Computer: scroll({'scroll_x': 0, 'scroll_y': 685, 'x': 633, 'y': 405})\n", - "INFO:agent.ComputerAgent:Computer: scroll({'scroll_x': 0, 'scroll_y': 685, 'x': 633, 'y': 405})\n", - " 18%|███████---------------------------------| 1334/7340 [45:13<203:35, 29.5 steps/min]2025-08-11 16:11:32,197 - agent.ComputerAgent - INFO - LLM processing started with 26 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 26 messages\n", - "\u001b[92m16:11:32 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:11:32 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 18%|███████---------------------------------| 1335/7340 [45:14<203:31, 29.5 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:11:33,951 - agent.ComputerAgent - INFO - LLM processing started with 10 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 10 messages\n", - "\u001b[92m16:11:33 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m16:11:33 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:11:35,279 - agent.ComputerAgent - INFO - Computer: type({'text': \"bash -lc 'cd ~/Desktop && ls -1 *.png'\"})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': \"bash -lc 'cd ~/Desktop && ls -1 *.png'\"})\n", - "2025-08-11 16:11:35,955 - agent.ComputerAgent - INFO - Computer: double_click({'x': 49, 'y': 431})\n", - "INFO:agent.ComputerAgent:Computer: double_click({'x': 49, 'y': 431})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:11:36 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/ffbf23fa-9bd6-4b26-befa-cb45d31fc4fa/invoke \"HTTP/1.1 200 OK\"\n", - " 18%|███████---------------------------------| 1335/7340 [45:18<203:47, 29.5 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:11:37,317 - agent.ComputerAgent - INFO - LLM processing started with 34 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 34 messages\n", - "\u001b[92m16:11:37 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m16:11:37 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:11:37,999 - agent.ComputerAgent - INFO - Computer: click({'x': 122, 'y': 176})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 122, 'y': 176})\n", - " 18%|███████---------------------------------| 1337/7340 [45:19<203:31, 29.5 steps/min]2025-08-11 16:11:38,642 - agent.ComputerAgent - INFO - LLM processing started with 26 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 26 messages\n", - "\u001b[92m16:11:38 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m16:11:39 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 18%|███████---------------------------------| 1338/7340 [45:21<203:26, 29.5 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m16:11:39 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:11:40,487 - agent.ComputerAgent - INFO - Computer: click({'x': 397, 'y': 624})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 397, 'y': 624})\n", - " 18%|███████---------------------------------| 1339/7340 [45:23<203:24, 29.5 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/0cad7a26-2224-4401-9a66-57daca76d380/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/79295f2f-2987-488c-b4b7-c968f71c7597/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 16:11:42,633 - agent.ComputerAgent - INFO - LLM processing started with 26 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 26 messages\n", - "\u001b[92m16:11:42 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/ca85c226-0c49-4084-b2bc-86bd540c8bce/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:11:43 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m16:11:43 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 18%|███████---------------------------------| 1339/7340 [45:25<203:35, 29.5 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m16:11:44 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:11:44 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:11:45,987 - agent.ComputerAgent - INFO - Computer: type({'text': 'find . -type f -perm -not -0644 -ls | head -n 20'})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': 'find . -type f -perm -not -0644 -ls | head -n 20'})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/cc2e38be-6768-4928-bfe5-d7f31cb68b24/invoke \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:11:46,670 - agent.ComputerAgent - INFO - Computer: click({'x': 249, 'y': 339})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 249, 'y': 339})\n", - "\u001b[92m16:11:46 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 18%|███████---------------------------------| 1339/7340 [45:28<203:47, 29.4 steps/min]\u001b[92m16:11:46 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:11:47,338 - agent.ComputerAgent - INFO - Computer: click({'x': 982, 'y': 741})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 982, 'y': 741})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:11:48 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/d8b3a739-de56-40fe-896f-831373c8ecee/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:11:48,711 - agent.ComputerAgent - INFO - Computer: click({'x': 1000, 'y': 739})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 1000, 'y': 739})\n", - " 18%|███████---------------------------------| 1341/7340 [45:30<203:34, 29.5 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:11:49,375 - agent.ComputerAgent - INFO - LLM processing started with 38 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 38 messages\n", - "\u001b[92m16:11:49 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m16:11:49 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:11:50,053 - agent.ComputerAgent - INFO - Computer: click({'x': 553, 'y': 275})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 553, 'y': 275})\n", - " 18%|███████---------------------------------| 1343/7340 [45:31<203:18, 29.5 steps/min]2025-08-11 16:11:50,708 - agent.ComputerAgent - INFO - LLM processing started with 36 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 36 messages\n", - "\u001b[92m16:11:50 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:11:51 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 18%|███████---------------------------------| 1344/7340 [45:33<203:14, 29.5 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:11:52,081 - agent.ComputerAgent - INFO - LLM processing started with 14 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 14 messages\n", - "\u001b[92m16:11:52 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m16:11:52 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:11:52,743 - agent.ComputerAgent - INFO - LLM processing started with 8 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 8 messages\n", - "\u001b[92m16:11:52 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 16:11:54,178 - agent.ComputerAgent - INFO - Computer: click({'x': 49, 'y': 53})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 49, 'y': 53})\n", - " 18%|███████---------------------------------| 1344/7340 [45:35<203:25, 29.5 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 16:11:55,482 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'ctrl+shift+v'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'ctrl+shift+v'})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/80299c20-3bcf-48b1-a471-299a1eda0a00/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/3b3e7fbd-8c02-45a6-bb3d-83c056398d3f/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:11:56 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/932fb6ee-8e77-41ca-8220-27e0c8783ced/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 18%|███████---------------------------------| 1345/7340 [45:37<203:23, 29.5 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:11:56,832 - agent.ComputerAgent - INFO - LLM processing started with 18 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 18 messages\n", - "\u001b[92m16:11:56 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m16:11:56 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:11:57,492 - agent.ComputerAgent - INFO - LLM processing started with 20 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 20 messages\n", - "\u001b[92m16:11:57 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 16:11:58,155 - agent.ComputerAgent - INFO - Computer: click({'x': 496, 'y': 256})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 496, 'y': 256})\n", - " 18%|███████---------------------------------| 1345/7340 [45:39<203:32, 29.5 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 16:11:59,512 - agent.ComputerAgent - INFO - Computer: click({'x': 188, 'y': 54})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 188, 'y': 54})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/89cdf329-a61d-4d69-9c6c-5d0ea35677b6/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 16:12:00,870 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'ctrl+f'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'ctrl+f'})\n", - " 18%|███████---------------------------------| 1346/7340 [45:42<203:33, 29.4 steps/min]2025-08-11 16:12:01,532 - agent.ComputerAgent - INFO - LLM processing started with 18 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 18 messages\n", - "\u001b[92m16:12:01 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 16:12:02,223 - agent.ComputerAgent - INFO - LLM processing started with 16 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 16 messages\n", - "\u001b[92m16:12:02 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/e8a299f4-d946-4970-b9a4-2503717de8ce/invoke \"HTTP/1.1 200 OK\"\n", - " 18%|███████---------------------------------| 1347/7340 [45:44<203:28, 29.5 steps/min]2025-08-11 16:12:02,913 - agent.ComputerAgent - INFO - LLM processing started with 22 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 22 messages\n", - "\u001b[92m16:12:02 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 16:12:03,573 - agent.ComputerAgent - INFO - LLM processing started with 20 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 20 messages\n", - "\u001b[92m16:12:03 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 18%|███████---------------------------------| 1347/7340 [45:45<203:34, 29.4 steps/min]2025-08-11 16:12:04,252 - agent.ComputerAgent - INFO - LLM processing started with 18 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 18 messages\n", - "\u001b[92m16:12:04 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 18%|███████---------------------------------| 1347/7340 [45:46<203:38, 29.4 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:12:05 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - " 18%|███████---------------------------------| 1347/7340 [45:47<203:43, 29.4 steps/min]\u001b[92m16:12:06 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:12:06,650 - agent.ComputerAgent - INFO - Computer: scroll({'scroll_x': 0, 'scroll_y': 677, 'x': 633, 'y': 362})\n", - "INFO:agent.ComputerAgent:Computer: scroll({'scroll_x': 0, 'scroll_y': 677, 'x': 633, 'y': 362})\n", - " 18%|███████---------------------------------| 1347/7340 [45:48<203:47, 29.4 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:12:07 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/b2656d0e-a6f4-4ecb-a099-cfe8471c4998/invoke \"HTTP/1.1 200 OK\"\n", - " 18%|███████---------------------------------| 1348/7340 [45:49<203:41, 29.4 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/fa4f593f-4977-4dc4-9238-0a67602a0900/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:12:08 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m16:12:08 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m16:12:09 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/c3f7029e-7bbd-43fb-bea4-c66cc9ae685d/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:12:09,759 - agent.ComputerAgent - INFO - Computer: click({'x': 296, 'y': 736})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 296, 'y': 736})\n", - " 18%|███████---------------------------------| 1348/7340 [45:51<203:50, 29.4 steps/min]\u001b[92m16:12:09 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:12:10,408 - agent.ComputerAgent - INFO - Computer: click({'x': 234, 'y': 97})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 234, 'y': 97})\n", - "\u001b[92m16:12:10 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:12:11,099 - agent.ComputerAgent - INFO - Computer: click({'x': 332, 'y': 162})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 332, 'y': 162})\n", - " 18%|███████---------------------------------| 1349/7340 [45:52<203:45, 29.4 steps/min]2025-08-11 16:12:11,772 - agent.ComputerAgent - INFO - LLM processing started with 42 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 42 messages\n", - "\u001b[92m16:12:11 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 16:12:12,423 - agent.ComputerAgent - INFO - LLM processing started with 28 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 28 messages\n", - "\u001b[92m16:12:12 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 18%|███████---------------------------------| 1351/7340 [45:54<203:29, 29.4 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/0cad7a26-2224-4401-9a66-57daca76d380/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 16:12:13,589 - agent.ComputerAgent - INFO - LLM processing started with 12 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 12 messages\n", - "\u001b[92m16:12:13 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:12:14 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 18%|███████---------------------------------| 1351/7340 [45:56<203:37, 29.4 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:12:14 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m16:12:15 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m16:12:16 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 18%|███████---------------------------------| 1351/7340 [45:58<203:46, 29.4 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m16:12:16 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:12:16,944 - agent.ComputerAgent - INFO - LLM processing started with 28 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 28 messages\n", - "\u001b[92m16:12:16 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:12:17,601 - agent.ComputerAgent - INFO - Computer: click({'x': 684, 'y': 41})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 684, 'y': 41})\n", - "\u001b[92m16:12:17 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:12:17 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m16:12:17 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 18%|███████---------------------------------| 1351/7340 [46:00<203:55, 29.4 steps/min]\u001b[92m16:12:18 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:12:18,978 - agent.ComputerAgent - INFO - Computer: click({'x': 122, 'y': 213})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 122, 'y': 213})\n", - "2025-08-11 16:12:19,629 - agent.ComputerAgent - INFO - Computer: click({'x': 244, 'y': 149})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 244, 'y': 149})\n", - "2025-08-11 16:12:20,317 - agent.ComputerAgent - INFO - Computer: click({'x': 397, 'y': 564})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 397, 'y': 564})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/80299c20-3bcf-48b1-a471-299a1eda0a00/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/fafe8f9a-bc46-42ad-b3ca-7190a64ab552/invoke \"HTTP/1.1 200 OK\"\n", - " 18%|███████---------------------------------| 1352/7340 [46:02<203:53, 29.4 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:12:20,972 - agent.ComputerAgent - INFO - LLM processing started with 20 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 20 messages\n", - "\u001b[92m16:12:20 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m16:12:20 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/ffbf23fa-9bd6-4b26-befa-cb45d31fc4fa/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:12:21,645 - agent.ComputerAgent - INFO - Computer: click({'x': 249, 'y': 339})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 249, 'y': 339})\n", - "2025-08-11 16:12:22,272 - agent.ComputerAgent - INFO - LLM processing started with 22 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 22 messages\n", - "\u001b[92m16:12:22 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 18%|███████---------------------------------| 1355/7340 [46:04<203:28, 29.4 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:12:22 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - " 18%|███████---------------------------------| 1356/7340 [46:05<203:22, 29.4 steps/min]2025-08-11 16:12:24,123 - agent.ComputerAgent - INFO - LLM processing started with 28 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 28 messages\n", - "\u001b[92m16:12:24 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m16:12:24 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:12:24,852 - agent.ComputerAgent - INFO - Computer: click({'x': 623, 'y': 359})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 623, 'y': 359})\n", - " 18%|███████---------------------------------| 1356/7340 [46:06<203:28, 29.4 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 16:12:26,220 - agent.ComputerAgent - INFO - Computer: type({'text': 'source=~/Desktop/dir1; target=~/Desktop/dir3; if [ -d \"$source\" ] && [ -d \"$target\" ]; then rsync -a -f\"+ */\" -f\"- *\" \"$source\" \"$target\"; echo \"Copied directory hierarchy.\"; else echo \"Source or target directory not found\"; fi'})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': 'source=~/Desktop/dir1; target=~/Desktop/dir3; if [ -d \"$source\" ] && [ -d \"$target\" ]; then rsync -a -f\"+ */\" -f\"- *\" \"$source\" \"$target\"; echo \"Copied directory hierarchy.\"; else echo \"Source or target directory not found\"; fi'})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:12:26 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 18%|███████---------------------------------| 1357/7340 [46:08<203:27, 29.4 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m16:12:27 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:12:28,062 - agent.ComputerAgent - INFO - Computer: click({'x': 72, 'y': 90})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 72, 'y': 90})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:12:28 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:12:30,106 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'enter'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'enter'})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/932fb6ee-8e77-41ca-8220-27e0c8783ced/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/cc2e38be-6768-4928-bfe5-d7f31cb68b24/invoke \"HTTP/1.1 200 OK\"\n", - " 19%|███████---------------------------------| 1358/7340 [46:11<203:29, 29.4 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/89cdf329-a61d-4d69-9c6c-5d0ea35677b6/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/d8b3a739-de56-40fe-896f-831373c8ecee/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:12:31,450 - agent.ComputerAgent - INFO - Computer: scroll({'scroll_y': 676, 'scroll_x': 0})\n", - "INFO:agent.ComputerAgent:Computer: scroll({'scroll_y': 676, 'scroll_x': 0})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/ca85c226-0c49-4084-b2bc-86bd540c8bce/invoke \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:12:31 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m16:12:32 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 19%|███████---------------------------------| 1361/7340 [46:13<203:05, 29.4 steps/min]2025-08-11 16:12:32,793 - agent.ComputerAgent - INFO - Computer: click({'x': 553, 'y': 280})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 553, 'y': 280})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m16:12:33 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:12:34,134 - agent.ComputerAgent - INFO - LLM processing started with 20 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 20 messages\n", - "\u001b[92m16:12:34 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/a23ddde7-5509-407d-af64-ea09807c1af1/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m16:12:34 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/39724bde-60dd-471d-ba25-1ac9b1405c76/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 19%|███████---------------------------------| 1361/7340 [46:15<203:14, 29.4 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:12:35,495 - agent.ComputerAgent - INFO - Computer: type({'text': \"bash -lc 'cd ~/Desktop && ls -1'}\"})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': \"bash -lc 'cd ~/Desktop && ls -1'}\"})\n", - "2025-08-11 16:12:36,154 - agent.ComputerAgent - INFO - Computer: click({'x': 263, 'y': 318})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 263, 'y': 318})\n", - "2025-08-11 16:12:36,816 - agent.ComputerAgent - INFO - LLM processing started with 40 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 40 messages\n", - "\u001b[92m16:12:36 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m16:12:36 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 19%|███████---------------------------------| 1362/7340 [46:18<203:15, 29.4 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:12:37,525 - agent.ComputerAgent - INFO - Computer: click({'x': 426, 'y': 257})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 426, 'y': 257})\n", - "2025-08-11 16:12:38,193 - agent.ComputerAgent - INFO - LLM processing started with 24 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 24 messages\n", - "\u001b[92m16:12:38 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - " 19%|███████---------------------------------| 1364/7340 [46:20<203:02, 29.4 steps/min]\u001b[92m16:12:38 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/0cad7a26-2224-4401-9a66-57daca76d380/invoke \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n" - ] - }, - { - "name": "stdout", - "output_type": "stream", - "text": [ - "No screenshot found, taking screenshot\n" - ] - }, - { - "name": "stderr", - "output_type": "stream", - "text": [ - "2025-08-11 16:12:39,517 - agent.ComputerAgent - INFO - LLM processing started with 31 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 31 messages\n", - "\u001b[92m16:12:39 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m16:12:39 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:12:40,192 - agent.ComputerAgent - INFO - Computer: click({'x': 526, 'y': 232})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 526, 'y': 232})\n", - "2025-08-11 16:12:40,843 - agent.ComputerAgent - INFO - LLM processing started with 10 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 10 messages\n", - "\u001b[92m16:12:40 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - " 19%|███████---------------------------------| 1365/7340 [46:23<203:03, 29.4 steps/min]\u001b[92m16:12:41 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:12:42,204 - agent.ComputerAgent - INFO - LLM processing started with 18 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 18 messages\n", - "\u001b[92m16:12:42 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m16:12:42 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 16:12:43,551 - agent.ComputerAgent - INFO - Computer: click({'x': 835, 'y': 36})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 835, 'y': 36})\n", - "2025-08-11 16:12:44,241 - agent.ComputerAgent - INFO - LLM processing started with 36 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 36 messages\n", - "\u001b[92m16:12:44 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 16:12:44,933 - agent.ComputerAgent - INFO - Computer: click({'x': 433, 'y': 635})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 433, 'y': 635})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 400 Bad Request\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/b2656d0e-a6f4-4ecb-a099-cfe8471c4998/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/3b3e7fbd-8c02-45a6-bb3d-83c056398d3f/invoke \"HTTP/1.1 200 OK\"\n", - " 19%|███████---------------------------------| 1367/7340 [46:26<202:56, 29.4 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/e8a299f4-d946-4970-b9a4-2503717de8ce/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/79295f2f-2987-488c-b4b7-c968f71c7597/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 16:12:45,580 - agent.ComputerAgent - INFO - LLM processing started with 40 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 40 messages\n", - "\u001b[92m16:12:45 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 16:12:46,215 - agent.ComputerAgent - INFO - LLM processing started with 22 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 22 messages\n", - "\u001b[92m16:12:46 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 19%|███████---------------------------------| 1369/7340 [46:28<202:40, 29.5 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/0cad7a26-2224-4401-9a66-57daca76d380/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 16:12:46,933 - agent.ComputerAgent - INFO - LLM processing started with 38 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 38 messages\n", - "\u001b[92m16:12:47 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 16:12:47,614 - agent.ComputerAgent - INFO - LLM processing started with 33 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 33 messages\n", - "\u001b[92m16:12:47 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/80299c20-3bcf-48b1-a471-299a1eda0a00/invoke \"HTTP/1.1 200 OK\"\n", - " 19%|███████---------------------------------| 1369/7340 [46:29<202:46, 29.4 steps/min]2025-08-11 16:12:48,293 - agent.ComputerAgent - INFO - LLM processing started with 24 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 24 messages\n", - "\u001b[92m16:12:48 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 16:12:48,973 - agent.ComputerAgent - INFO - LLM processing started with 20 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 20 messages\n", - "\u001b[92m16:12:49 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 19%|███████---------------------------------| 1369/7340 [46:30<202:52, 29.4 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 16:12:50,174 - agent.ComputerAgent - INFO - LLM processing started with 14 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 14 messages\n", - "\u001b[92m16:12:50 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/fafe8f9a-bc46-42ad-b3ca-7190a64ab552/invoke \"HTTP/1.1 200 OK\"\n", - " 19%|███████---------------------------------| 1369/7340 [46:31<202:57, 29.4 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/c3f7029e-7bbd-43fb-bea4-c66cc9ae685d/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 16:12:51,233 - agent.ComputerAgent - INFO - LLM processing started with 22 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 22 messages\n", - "\u001b[92m16:12:51 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/89cdf329-a61d-4d69-9c6c-5d0ea35677b6/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 16:12:52,561 - agent.ComputerAgent - INFO - Computer: type({'text': 'file1'})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': 'file1'})\n", - " 19%|███████---------------------------------| 1369/7340 [46:34<203:07, 29.4 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:12:53 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 400 Bad Request\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/fa4f593f-4977-4dc4-9238-0a67602a0900/invoke \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:12:53,913 - agent.ComputerAgent - INFO - LLM processing started with 22 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 22 messages\n", - "\u001b[92m16:12:53 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 16:12:55,675 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'ctrl+,'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'ctrl+,'})\n", - " 19%|███████---------------------------------| 1371/7340 [46:37<202:59, 29.4 steps/min]\u001b[92m16:12:55 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:12:56,375 - agent.ComputerAgent - INFO - Computer: click({'x': 304, 'y': 735})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 304, 'y': 735})\n", - "2025-08-11 16:12:57,004 - agent.ComputerAgent - INFO - LLM processing started with 12 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 12 messages\n", - "\u001b[92m16:12:57 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 19%|███████---------------------------------| 1371/7340 [46:38<203:05, 29.4 steps/min]2025-08-11 16:12:57,682 - agent.ComputerAgent - INFO - LLM processing started with 30 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 30 messages\n", - "\u001b[92m16:12:57 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 19%|███████---------------------------------| 1372/7340 [46:39<202:58, 29.4 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 16:13:00,048 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'ctrl+v'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'ctrl+v'})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/0cad7a26-2224-4401-9a66-57daca76d380/invoke \"HTTP/1.1 200 OK\"\n", - " 19%|███████---------------------------------| 1372/7340 [46:41<203:07, 29.4 steps/min]2025-08-11 16:13:00,705 - agent.ComputerAgent - INFO - LLM processing started with 35 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 35 messages\n", - "\u001b[92m16:13:00 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 16:13:01,364 - agent.ComputerAgent - INFO - LLM processing started with 24 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 24 messages\n", - "\u001b[92m16:13:01 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 19%|███████---------------------------------| 1372/7340 [46:45<203:22, 29.3 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/ca85c226-0c49-4084-b2bc-86bd540c8bce/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/fa4f593f-4977-4dc4-9238-0a67602a0900/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:13:04 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 400 Bad Request\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 19%|███████---------------------------------| 1374/7340 [46:46<203:05, 29.4 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:13:05,255 - agent.ComputerAgent - INFO - LLM processing started with 20 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 20 messages\n", - "\u001b[92m16:13:05 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m16:13:05 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/fa4f593f-4977-4dc4-9238-0a67602a0900/close \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 16:13:06,551 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'enter'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'enter'})\n", - "2025-08-11 16:13:07,203 - agent.ComputerAgent - INFO - Computer: click({'x': 232, 'y': 97})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 232, 'y': 97})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/ffbf23fa-9bd6-4b26-befa-cb45d31fc4fa/invoke \"HTTP/1.1 200 OK\"\n", - " 19%|███████---------------------------------| 1376/7340 [46:49<202:59, 29.4 steps/min]2025-08-11 16:13:08,546 - agent.ComputerAgent - INFO - LLM processing started with 30 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 30 messages\n", - "\u001b[92m16:13:08 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:13:09 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: GET https://orchestration.hud.so/hud-gym/api/v1/gyms/OSWorld-Ubuntu \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/0cad7a26-2224-4401-9a66-57daca76d380/invoke \"HTTP/1.1 200 OK\"\n", - " 19%|███████---------------------------------| 1376/7340 [46:51<203:04, 29.4 steps/min]2025-08-11 16:13:09,882 - agent.ComputerAgent - INFO - LLM processing started with 37 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 37 messages\n", - "\u001b[92m16:13:09 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 19%|███████---------------------------------| 1376/7340 [46:52<203:08, 29.4 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:13:11 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:accelerate.utils.modeling:We will use 90% of the memory on device 0 for storing the model, and 10% for the buffer to avoid OOM. You can set `max_memory` in to a higher value to use more memory (at your own risk).\n", - "Loading checkpoint shards: 0%| | 0/4 [00:00'})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': ''})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/a4e9c5c3-fa17-4f05-8383-03a3cb3c1fba/invoke \"HTTP/1.1 200 OK\"\n", - " 23%|█████████-------------------------------| 1664/7340 [58:10<198:24, 28.6 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/4813e5e3-be12-40e2-9cc0-d5be0ad320cf/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:24:29 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m16:24:29 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/c062c21a-1b89-4117-86d3-d763f8af4cbd/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/7a0a74ba-160b-41ee-a6d2-6dc61c143d94/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 400 Bad Request\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:24:29 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m16:24:30 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 23%|█████████-------------------------------| 1684/7340 [58:12<195:28, 28.9 steps/min]2025-08-11 16:24:31,005 - agent.ComputerAgent - INFO - Computer: click({'x': 369, 'y': 564})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 369, 'y': 564})\n", - "\u001b[92m16:24:31 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m16:24:31 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:24:31,655 - agent.ComputerAgent - INFO - Computer: click({'x': 842, 'y': 571})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 842, 'y': 571})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/create_environment \"HTTP/1.1 200 OK\"\n", - "2025-08-11 16:24:32,328 - agent.ComputerAgent - INFO - LLM processing started with 26 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 26 messages\n", - "\u001b[92m16:24:32 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 16:24:32,999 - agent.ComputerAgent - INFO - LLM processing started with 16 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 16 messages\n", - "\u001b[92m16:24:33 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m16:24:33 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m16:24:33 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 23%|█████████-------------------------------| 1684/7340 [58:14<195:37, 28.9 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:24:33 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:24:34,996 - agent.ComputerAgent - INFO - Computer: type({'text': '3'})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': '3'})\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:24:35,673 - agent.ComputerAgent - INFO - Computer: click({'button': 'right', 'x': 987, 'y': 658})\n", - "INFO:agent.ComputerAgent:Computer: click({'button': 'right', 'x': 987, 'y': 658})\n", - "2025-08-11 16:24:36,338 - agent.ComputerAgent - INFO - LLM processing started with 6 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 6 messages\n", - "\u001b[92m16:24:36 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 23%|█████████-------------------------------| 1686/7340 [58:18<195:30, 28.9 steps/min]\u001b[92m16:24:36 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:24:37,031 - agent.ComputerAgent - INFO - Computer: drag({'path': [{'x': 280, 'y': 375}, {'x': 802, 'y': 446}]})\n", - "INFO:agent.ComputerAgent:Computer: drag({'path': [{'x': 280, 'y': 375}, {'x': 802, 'y': 446}]})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:24:37 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m16:24:37 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 23%|█████████-------------------------------| 1688/7340 [58:19<195:17, 28.9 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:24:38,396 - agent.ComputerAgent - INFO - Computer: click({'x': 60, 'y': 35})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 60, 'y': 35})\n", - "\u001b[92m16:24:38 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/e8a299f4-d946-4970-b9a4-2503717de8ce/close \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/7bc07116-76e3-42fb-a0e3-a2273a5caa64/close \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/d8b3a739-de56-40fe-896f-831373c8ecee/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 16:24:39,048 - agent.ComputerAgent - INFO - Computer: click({'x': 478, 'y': 256})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 478, 'y': 256})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/a4e9c5c3-fa17-4f05-8383-03a3cb3c1fba/close \"HTTP/1.1 200 OK\"\n", - " 23%|█████████-------------------------------| 1693/7340 [58:22<194:41, 29.0 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/3ad517be-7b27-424d-b632-3ba6ff1a1e71/invoke \"HTTP/1.1 200 OK\"\n", - " 23%|█████████-------------------------------| 1695/7340 [58:23<194:26, 29.0 steps/min]INFO:httpx:HTTP Request: GET https://orchestration.hud.so/hud-gym/api/v1/gyms/OSWorld-Ubuntu \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: GET https://orchestration.hud.so/hud-gym/api/v1/gyms/OSWorld-Ubuntu \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: GET https://orchestration.hud.so/hud-gym/api/v1/gyms/OSWorld-Ubuntu \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/d8b3a739-de56-40fe-896f-831373c8ecee/close \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/3980166d-0a7d-4a58-a915-07dbe8b607bb/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/b6b06a1a-197c-499e-a884-cc6bce509fa3/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - " 23%|█████████-------------------------------| 1695/7340 [58:25<194:33, 29.0 steps/min]\u001b[92m16:24:43 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/9dca7e41-642b-4cca-8758-834cef0e844c/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/5da0c259-034b-4ba2-9e95-9d4ae99c7475/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:accelerate.utils.modeling:We will use 90% of the memory on device 0 for storing the model, and 10% for the buffer to avoid OOM. You can set `max_memory` in to a higher value to use more memory (at your own risk).\n", - "Loading checkpoint shards: 0%| | 0/4 [00:00\n", - "INFO:accelerate.utils.modeling:We will use 90% of the memory on device 0 for storing the model, and 10% for the buffer to avoid OOM. You can set `max_memory` in to a higher value to use more memory (at your own risk).\n", - " 26%|██████████------------------------------| 1938/7340 [66:27<185:15, 29.2 steps/min]\u001b[92m16:32:46 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/8956c64b-871b-43e2-84de-047c8ce2a839/invoke \"HTTP/1.1 200 OK\"\n", - "Loading checkpoint shards: 100%|██████████| 4/4 [00:05<00:00, 1.48s/it]\n", - "2025-08-11 16:32:46,573 - agent.ComputerAgent - INFO - LLM processing started with 26 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 26 messages\n", - "\u001b[92m16:32:46 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 26%|██████████------------------------------| 1939/7340 [66:29<185:13, 29.2 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/e7117b51-399c-45d8-88a1-c54a00b2bc38/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 16:32:48,835 - agent.ComputerAgent - INFO - LLM processing started with 16 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 16 messages\n", - "\u001b[92m16:32:48 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 26%|██████████------------------------------| 1939/7340 [66:32<185:19, 29.1 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:32:51 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Loading checkpoint shards: 100%|██████████| 4/4 [00:05<00:00, 1.38s/it]29.1 steps/min]\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:32:51 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:32:53,100 - agent.ComputerAgent - INFO - Computer: type({'text': '\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08LARS Resources (Backup)'})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': '\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08\\x08LARS Resources (Backup)'})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:32:54,407 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'ctrl+l'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'ctrl+l'})\n", - " 26%|██████████------------------------------| 1939/7340 [66:36<185:31, 29.1 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:32:54 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m16:32:54 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m16:32:54 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m16:32:55 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:32:55,743 - agent.ComputerAgent - INFO - Computer: click({'x': 1009, 'y': 101})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 1009, 'y': 101})\n", - "2025-08-11 16:32:56,380 - agent.ComputerAgent - INFO - Computer: click({'x': 46, 'y': 527})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 46, 'y': 527})\n", - "2025-08-11 16:32:57,051 - agent.ComputerAgent - INFO - Computer: click({'x': 72, 'y': 244})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 72, 'y': 244})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:32:57 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:32:58,377 - agent.ComputerAgent - INFO - Computer: screenshot({})\n", - "INFO:agent.ComputerAgent:Computer: screenshot({})\n", - "2025-08-11 16:32:59,058 - agent.ComputerAgent - INFO - LLM processing started with 42 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 42 messages\n", - "\u001b[92m16:32:59 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 16:32:59,707 - agent.ComputerAgent - INFO - Computer: click({'x': 693, 'y': 698})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 693, 'y': 698})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:32:59 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:33:00 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:33:01 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 26%|██████████------------------------------| 1940/7340 [66:43<185:43, 29.1 steps/min]\u001b[92m16:33:01 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:33:02,296 - agent.ComputerAgent - INFO - Computer: click({'x': 70, 'y': 77})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 70, 'y': 77})\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - " 26%|██████████------------------------------| 1945/7340 [66:44<185:07, 29.1 steps/min]\u001b[92m16:33:02 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:33:03,527 - agent.ComputerAgent - INFO - Computer: click({'x': 635, 'y': 468})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 635, 'y': 468})\n", - "\u001b[92m16:33:03 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:33:04,208 - agent.ComputerAgent - INFO - Computer: scroll({'scroll_y': -628, 'scroll_x': 0, 'x': 526, 'y': 463})\n", - "INFO:agent.ComputerAgent:Computer: scroll({'scroll_y': -628, 'scroll_x': 0, 'x': 526, 'y': 463})\n", - "\u001b[92m16:33:04 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/4adb2bbf-d6e6-4d15-9e9a-c199cf02d5d6/invoke \"HTTP/1.1 200 OK\"\n", - " 27%|██████████------------------------------| 1946/7340 [66:45<185:03, 29.1 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n" - ] - }, - { - "name": "stdout", - "output_type": "stream", - "text": [ - "No screenshot found, taking screenshot\n" - ] - }, - { - "name": "stderr", - "output_type": "stream", - "text": [ - "2025-08-11 16:33:04,854 - agent.ComputerAgent - INFO - LLM processing started with 8 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 8 messages\n", - "\u001b[92m16:33:04 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 16:33:05,505 - agent.ComputerAgent - INFO - Computer: click({'x': 969, 'y': 169})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 969, 'y': 169})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:33:06 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 27%|██████████------------------------------| 1948/7340 [66:47<184:53, 29.2 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:33:06 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m16:33:06 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:33:07,437 - agent.ComputerAgent - INFO - Computer: click({'x': 87, 'y': 181})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 87, 'y': 181})\n", - "\u001b[92m16:33:07 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/3980166d-0a7d-4a58-a915-07dbe8b607bb/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/f83df7e3-6ab0-404e-9745-09768e42b6fb/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/6d8a38cc-c8f6-484c-9a6d-e6c404b2c7f9/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/4813e5e3-be12-40e2-9cc0-d5be0ad320cf/invoke \"HTTP/1.1 200 OK\"\n", - " 27%|██████████------------------------------| 1949/7340 [66:49<184:49, 29.2 steps/min]2025-08-11 16:33:08,125 - agent.ComputerAgent - INFO - Computer: click({'x': 76, 'y': 321})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 76, 'y': 321})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/a3ea8855-19d9-4e10-8208-fd9e060997e3/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 16:33:08,772 - agent.ComputerAgent - INFO - LLM processing started with 32 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 32 messages\n", - "\u001b[92m16:33:08 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 16:33:09,435 - agent.ComputerAgent - INFO - LLM processing started with 14 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 14 messages\n", - "\u001b[92m16:33:09 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 16:33:10,071 - agent.ComputerAgent - INFO - LLM processing started with 28 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 28 messages\n", - "\u001b[92m16:33:10 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 27%|██████████------------------------------| 1950/7340 [66:51<184:49, 29.2 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 16:33:11,418 - agent.ComputerAgent - INFO - Computer: type({'text': 'orig=$(find . -path ./fails -prune -o -type f -name \"*failed.ipynb\" -print | wc -l); copied=$(find ./fails -type f -name \"*failed.ipynb\" -print | wc -l); echo \"orig=$orig copied=$copied\"'})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': 'orig=$(find . -path ./fails -prune -o -type f -name \"*failed.ipynb\" -print | wc -l); copied=$(find ./fails -type f -name \"*failed.ipynb\" -print | wc -l); echo \"orig=$orig copied=$copied\"'})\n", - "2025-08-11 16:33:12,446 - agent.ComputerAgent - INFO - LLM processing started with 36 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 36 messages\n", - "\u001b[92m16:33:12 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/ea9e43cc-3d54-4c89-bb53-a189a3ae9a25/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/d0000302-258b-4660-9baa-e149c2ad83fd/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/b2ca79e3-4425-4cd4-a9dd-42e2431eb008/invoke \"HTTP/1.1 200 OK\"\n", - " 27%|██████████------------------------------| 1951/7340 [66:54<184:48, 29.2 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 16:33:14,195 - agent.ComputerAgent - INFO - Computer: type({'text': 'sudo apt-get update -y && sudo apt-get install -y steghide binwalk exiftool ffmpeg\\n'})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': 'sudo apt-get update -y && sudo apt-get install -y steghide binwalk exiftool ffmpeg\\n'})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/e7117b51-399c-45d8-88a1-c54a00b2bc38/invoke \"HTTP/1.1 200 OK\"\n", - " 27%|██████████------------------------------| 1952/7340 [66:55<184:44, 29.2 steps/min]2025-08-11 16:33:14,816 - agent.ComputerAgent - INFO - LLM processing started with 18 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 18 messages\n", - "\u001b[92m16:33:14 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 16:33:15,505 - agent.ComputerAgent - INFO - LLM processing started with 26 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 26 messages\n", - "\u001b[92m16:33:15 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 16:33:16,172 - agent.ComputerAgent - INFO - LLM processing started with 18 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 18 messages\n", - "\u001b[92m16:33:16 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v1/environments/6b741091-faa0-4d97-9592-0dc410b6cc53/reset \"HTTP/1.1 200 OK\"\n", - " 27%|██████████------------------------------| 1953/7340 [66:57<184:42, 29.2 steps/min]2025-08-11 16:33:16,865 - agent.ComputerAgent - INFO - LLM processing started with 8 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 8 messages\n", - "\u001b[92m16:33:16 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 27%|██████████------------------------------| 1953/7340 [66:58<184:45, 29.2 steps/min]2025-08-11 16:33:17,496 - agent.ComputerAgent - INFO - LLM processing started with 8 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 8 messages\n", - "\u001b[92m16:33:17 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 27%|██████████------------------------------| 1953/7340 [67:01<184:53, 29.1 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/6b741091-faa0-4d97-9592-0dc410b6cc53/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:33:21 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/81398d20-3c85-489b-9abc-2af244ec1feb/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/7a0a74ba-160b-41ee-a6d2-6dc61c143d94/invoke \"HTTP/1.1 200 OK\"\n", - " 27%|██████████------------------------------| 1953/7340 [67:03<184:58, 29.1 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:33:22,384 - agent.ComputerAgent - INFO - LLM processing started with 16 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 16 messages\n", - "\u001b[92m16:33:22 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m16:33:22 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/8956c64b-871b-43e2-84de-047c8ce2a839/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m16:33:23 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:33:23,684 - agent.ComputerAgent - INFO - Computer: click({'x': 237, 'y': 95})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 237, 'y': 95})\n", - " 27%|██████████------------------------------| 1953/7340 [67:05<185:03, 29.1 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:33:24,329 - agent.ComputerAgent - INFO - LLM processing started with 38 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 38 messages\n", - "\u001b[92m16:33:24 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/488d7653-4f2d-4576-85c7-d87dc7a875ef/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m16:33:24 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:33:25 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m16:33:25 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m16:33:26 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:33:27,033 - agent.ComputerAgent - INFO - Computer: click({'x': 592, 'y': 568})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 592, 'y': 568})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m16:33:27 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 27%|██████████------------------------------| 1954/7340 [67:09<185:06, 29.1 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m16:33:27 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:33:28,365 - agent.ComputerAgent - INFO - Computer: click({'x': 664, 'y': 213})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 664, 'y': 213})\n", - "\u001b[92m16:33:28 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:33:29,040 - agent.ComputerAgent - INFO - Computer: click({'x': 489, 'y': 427})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 489, 'y': 427})\n", - "\u001b[92m16:33:29 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m16:33:29 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 27%|██████████------------------------------| 1955/7340 [67:10<185:02, 29.1 steps/min]2025-08-11 16:33:29,694 - agent.ComputerAgent - INFO - Computer: click({'x': 83, 'y': 139})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 83, 'y': 139})\n", - "2025-08-11 16:33:30,372 - agent.ComputerAgent - INFO - Computer: scroll({'scroll_y': -658, 'scroll_x': 0, 'x': 526, 'y': 432})\n", - "INFO:agent.ComputerAgent:Computer: scroll({'scroll_y': -658, 'scroll_x': 0, 'x': 526, 'y': 432})\n", - " 27%|██████████------------------------------| 1957/7340 [67:12<184:50, 29.1 steps/min]2025-08-11 16:33:31,077 - agent.ComputerAgent - INFO - LLM processing started with 28 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 28 messages\n", - "\u001b[92m16:33:31 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 16:33:31,780 - agent.ComputerAgent - INFO - LLM processing started with 1 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 1 messages\n", - "\u001b[92m16:33:31 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 16:33:33,106 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'ctrl+a'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'ctrl+a'})\n", - " 27%|██████████------------------------------| 1959/7340 [67:14<184:42, 29.1 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:33:33 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:33:34,429 - agent.ComputerAgent - INFO - LLM processing started with 20 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 20 messages\n", - "\u001b[92m16:33:34 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 27%|██████████------------------------------| 1959/7340 [67:16<184:46, 29.1 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:33:35,086 - agent.ComputerAgent - INFO - LLM processing started with 24 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 24 messages\n", - "\u001b[92m16:33:35 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m16:33:35 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/6d8a38cc-c8f6-484c-9a6d-e6c404b2c7f9/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/f83df7e3-6ab0-404e-9745-09768e42b6fb/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/4adb2bbf-d6e6-4d15-9e9a-c199cf02d5d6/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 16:33:35,779 - agent.ComputerAgent - INFO - Computer: click({'x': 86, 'y': 73})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 86, 'y': 73})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/4813e5e3-be12-40e2-9cc0-d5be0ad320cf/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/485267e4-f348-45f0-a08d-1d1f28a01f1d/invoke \"HTTP/1.1 200 OK\"\n", - " 27%|██████████------------------------------| 1959/7340 [67:17<184:50, 29.1 steps/min]2025-08-11 16:33:36,470 - agent.ComputerAgent - INFO - LLM processing started with 16 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 16 messages\n", - "\u001b[92m16:33:36 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 16:33:37,105 - agent.ComputerAgent - INFO - LLM processing started with 30 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 30 messages\n", - "\u001b[92m16:33:37 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - " 27%|██████████------------------------------| 1960/7340 [67:19<184:48, 29.1 steps/min]\u001b[92m16:33:37 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:33:38,439 - agent.ComputerAgent - INFO - LLM processing started with 10 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 10 messages\n", - "\u001b[92m16:33:38 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:33:39,107 - agent.ComputerAgent - INFO - LLM processing started with 38 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 38 messages\n", - "\u001b[92m16:33:39 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 27%|██████████------------------------------| 1960/7340 [67:20<184:51, 29.1 steps/min]\u001b[92m16:33:39 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:33:39,749 - agent.ComputerAgent - INFO - Computer: click({'x': 715, 'y': 627})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 715, 'y': 627})\n", - " 27%|██████████------------------------------| 1960/7340 [67:21<184:54, 29.1 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/485267e4-f348-45f0-a08d-1d1f28a01f1d/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/e7117b51-399c-45d8-88a1-c54a00b2bc38/invoke \"HTTP/1.1 200 OK\"\n", - " 27%|██████████------------------------------| 1962/7340 [67:22<184:41, 29.1 steps/min]2025-08-11 16:33:41,945 - agent.ComputerAgent - INFO - LLM processing started with 20 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 20 messages\n", - "\u001b[92m16:33:41 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/485267e4-f348-45f0-a08d-1d1f28a01f1d/close \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 27%|██████████------------------------------| 1962/7340 [67:24<184:45, 29.1 steps/min]INFO:httpx:HTTP Request: GET https://orchestration.hud.so/hud-gym/api/v1/gyms/OSWorld-Ubuntu \"HTTP/1.1 200 OK\"\n", - " 27%|██████████------------------------------| 1962/7340 [67:25<184:48, 29.1 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:33:44 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/ea9e43cc-3d54-4c89-bb53-a189a3ae9a25/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:accelerate.utils.modeling:We will use 90% of the memory on device 0 for storing the model, and 10% for the buffer to avoid OOM. You can set `max_memory` in to a higher value to use more memory (at your own risk).\n", - "Loading checkpoint shards: 0%| | 0/4 [00:00\\n2
\\n3
'})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': '1
\\n2
\\n3
'})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/2c254802-788e-4b4b-98dc-68cd2c6bcce4/invoke \"HTTP/1.1 200 OK\"\n", - " 30%|███████████-----------------------------| 2167/7340 [73:25<175:17, 29.5 steps/min]2025-08-11 16:39:45,389 - agent.ComputerAgent - INFO - LLM processing started with 22 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 22 messages\n", - "\u001b[92m16:39:45 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 30%|███████████-----------------------------| 2168/7340 [73:27<175:13, 29.5 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/8956c64b-871b-43e2-84de-047c8ce2a839/invoke \"HTTP/1.1 200 OK\"\n", - " 30%|███████████-----------------------------| 2168/7340 [73:28<175:16, 29.5 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 16:39:48,220 - agent.ComputerAgent - INFO - Computer: type({'text': 'Manchester, GB'})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': 'Manchester, GB'})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/8956c64b-871b-43e2-84de-047c8ce2a839/close \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - " 30%|███████████-----------------------------| 2168/7340 [73:30<175:22, 29.5 steps/min]\u001b[92m16:39:48 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - " 30%|███████████-----------------------------| 2169/7340 [73:31<175:17, 29.5 steps/min]\u001b[92m16:39:50 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:39:50,735 - agent.ComputerAgent - INFO - Agent: Taking a screenshot to see the current computer screen.\n", - "INFO:agent.ComputerAgent:Agent: Taking a screenshot to see the current computer screen.\n", - "2025-08-11 16:39:50,735 - agent.ComputerAgent - INFO - Computer: click({'x': 256, 'y': 173})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 256, 'y': 173})\n", - "INFO:httpx:HTTP Request: GET https://orchestration.hud.so/hud-gym/api/v1/gyms/OSWorld-Ubuntu \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/71840850-9565-4ed2-8fa2-e4f2ba6ec6a9/invoke \"HTTP/1.1 200 OK\"\n", - " 30%|███████████-----------------------------| 2169/7340 [73:32<175:20, 29.5 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:39:51 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:accelerate.utils.modeling:We will use 90% of the memory on device 0 for storing the model, and 10% for the buffer to avoid OOM. You can set `max_memory` in to a higher value to use more memory (at your own risk).\n", - "Loading checkpoint shards: 0%| | 0/4 [00:00> ~/.vimrc && echo Done\\''})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': 'bash -lc \\'printf \"\\\\n\\\\\" Enable absolute line numbers by default\\\\nset number\\\\n\" >> ~/.vimrc && echo Done\\''})\n", - " 31%|████████████----------------------------| 2276/7340 [77:50<173:11, 29.2 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/0803e2c2-9de2-40ff-93da-cb49f156cbba/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/029e3386-8f59-44a0-9dba-b14a4d9825ef/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/7e0b0038-3a97-4d93-8c5c-154cc0b95af9/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/6195bb79-4eff-4d3b-8b67-f28a4e6a73fa/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/74442f45-62aa-40d1-9499-ea3e8e0a4c18/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 16:44:10,219 - agent.ComputerAgent - INFO - LLM processing started with 30 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 30 messages\n", - "\u001b[92m16:44:10 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/a8ddfa4f-6f9b-4ad8-b763-1881394c9926/invoke \"HTTP/1.1 200 OK\"\n", - " 31%|████████████----------------------------| 2277/7340 [77:52<173:08, 29.2 steps/min]2025-08-11 16:44:10,891 - agent.ComputerAgent - INFO - LLM processing started with 16 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 16 messages\n", - "\u001b[92m16:44:10 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 16:44:11,562 - agent.ComputerAgent - INFO - LLM processing started with 28 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 28 messages\n", - "\u001b[92m16:44:11 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 31%|████████████----------------------------| 2277/7340 [77:53<173:11, 29.2 steps/min]2025-08-11 16:44:12,212 - agent.ComputerAgent - INFO - LLM processing started with 20 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 20 messages\n", - "\u001b[92m16:44:12 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:44:12 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:44:14,267 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'ctrl+a'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'ctrl+a'})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/475030e1-7ae3-4ced-91fb-2221d956a2ad/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 400 Bad Request\"\n", - "\u001b[92m16:44:14 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 16:44:16,173 - agent.ComputerAgent - INFO - Agent: The heading is now center-aligned and the document has been saved. Task completed.\n", - "INFO:agent.ComputerAgent:Agent: The heading is now center-aligned and the document has been saved. Task completed.\n", - "2025-08-11 16:44:16,851 - agent.ComputerAgent - INFO - Total usage:\n", - " - completion_tokens: 218\n", - " - prompt_tokens: 4658\n", - " - total_tokens: 4876\n", - " - completion_tokens_details:\n", - " - accepted_prediction_tokens: 0\n", - " - audio_tokens: 0\n", - " - reasoning_tokens: 192\n", - " - rejected_prediction_tokens: 0\n", - " - prompt_tokens_details:\n", - " - audio_tokens: 0\n", - " - cached_tokens: 0\n", - " - response_cost: $0.0080\n", - "INFO:agent.ComputerAgent:Total usage:\n", - " - completion_tokens: 218\n", - " - prompt_tokens: 4658\n", - " - total_tokens: 4876\n", - " - completion_tokens_details:\n", - " - accepted_prediction_tokens: 0\n", - " - audio_tokens: 0\n", - " - reasoning_tokens: 192\n", - " - rejected_prediction_tokens: 0\n", - " - prompt_tokens_details:\n", - " - audio_tokens: 0\n", - " - cached_tokens: 0\n", - " - response_cost: $0.0080\n", - " 31%|████████████----------------------------| 2279/7340 [77:58<173:09, 29.2 steps/min]2025-08-11 16:44:17,521 - agent.ComputerAgent - INFO - LLM processing started with 12 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 12 messages\n", - "\u001b[92m16:44:17 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 16:44:18,174 - agent.ComputerAgent - INFO - LLM processing started with 12 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 12 messages\n", - "\u001b[92m16:44:18 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:44:18 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m16:44:18 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:44:19,530 - agent.ComputerAgent - INFO - LLM processing started with 32 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 32 messages\n", - "\u001b[92m16:44:19 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - " 31%|████████████----------------------------| 2279/7340 [78:01<173:17, 29.2 steps/min]\u001b[92m16:44:20 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:44:20,843 - agent.ComputerAgent - INFO - Computer: click({'x': 1008, 'y': 193})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 1008, 'y': 193})\n", - "\u001b[92m16:44:20 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m16:44:21 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:44:22,141 - agent.ComputerAgent - INFO - Computer: click({'x': 776, 'y': 643})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 776, 'y': 643})\n", - "2025-08-11 16:44:22,791 - agent.ComputerAgent - INFO - LLM processing started with 20 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 20 messages\n", - " 31%|████████████----------------------------| 2279/7340 [78:04<173:22, 29.2 steps/min]\u001b[92m16:44:22 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m16:44:22 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m16:44:22 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:44:23,485 - agent.ComputerAgent - INFO - Computer: click({'x': 977, 'y': 16})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 977, 'y': 16})\n", - "2025-08-11 16:44:24,132 - agent.ComputerAgent - INFO - Computer: click({'x': 111, 'y': 162})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 111, 'y': 162})\n", - "\u001b[92m16:44:24 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 31%|████████████----------------------------| 2281/7340 [78:05<173:12, 29.2 steps/min]2025-08-11 16:44:24,809 - agent.ComputerAgent - INFO - Computer: click({'x': 625, 'y': 248})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 625, 'y': 248})\n", - " 31%|████████████----------------------------| 2283/7340 [78:06<173:01, 29.2 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/ed3b76a0-7eeb-473c-8da1-310545515f47/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 16:44:26,462 - agent.ComputerAgent - INFO - LLM processing started with 39 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 39 messages\n", - "\u001b[92m16:44:26 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/5180ec6f-26a5-4ab4-8ca3-87f128083da1/invoke \"HTTP/1.1 200 OK\"\n", - " 31%|████████████----------------------------| 2284/7340 [78:08<172:58, 29.2 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:44:27 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:44:28,788 - agent.ComputerAgent - INFO - Computer: type({'text': 'none.png'})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': 'none.png'})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:44:29 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 31%|████████████----------------------------| 2284/7340 [78:11<173:06, 29.2 steps/min]\u001b[92m16:44:30 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:44:30 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:44:30 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m16:44:31 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/ea9e43cc-3d54-4c89-bb53-a189a3ae9a25/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/ff8b808f-c3a6-4979-8f9a-c6a25905116c/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/a46ee6f6-d167-47c4-ad83-e16b88450253/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/a39ee9df-d3ba-456a-95cf-3a11a826583b/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 16:44:32,685 - agent.ComputerAgent - INFO - Computer: type({'text': 'pkill -f soffice.bin\\n'})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': 'pkill -f soffice.bin\\n'})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/3e4ea7d7-21a2-4b07-abd4-a3e280e44e0b/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 16:44:33,322 - agent.ComputerAgent - INFO - Computer: click({'x': 982, 'y': 632})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 982, 'y': 632})\n", - "\u001b[92m16:44:33 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m16:44:33 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:44:34,677 - agent.ComputerAgent - INFO - Agent: All .jpg files from the photos directory (recursively) have been copied into cpjpg on the Desktop. Verified: 4 .jpg files found and 4 copied.\n", - "\n", - "Task completed\n", - "INFO:agent.ComputerAgent:Agent: All .jpg files from the photos directory (recursively) have been copied into cpjpg on the Desktop. Verified: 4 .jpg files found and 4 copied.\n", - "\n", - "Task completed\n", - "2025-08-11 16:44:35,306 - agent.ComputerAgent - INFO - Total usage:\n", - " - completion_tokens: 494\n", - " - prompt_tokens: 8268\n", - " - total_tokens: 8762\n", - " - completion_tokens_details:\n", - " - accepted_prediction_tokens: 0\n", - " - audio_tokens: 0\n", - " - reasoning_tokens: 448\n", - " - rejected_prediction_tokens: 0\n", - " - prompt_tokens_details:\n", - " - audio_tokens: 0\n", - " - cached_tokens: 0\n", - " - response_cost: $0.0153\n", - "INFO:agent.ComputerAgent:Total usage:\n", - " - completion_tokens: 494\n", - " - prompt_tokens: 8268\n", - " - total_tokens: 8762\n", - " - completion_tokens_details:\n", - " - accepted_prediction_tokens: 0\n", - " - audio_tokens: 0\n", - " - reasoning_tokens: 448\n", - " - rejected_prediction_tokens: 0\n", - " - prompt_tokens_details:\n", - " - audio_tokens: 0\n", - " - cached_tokens: 0\n", - " - response_cost: $0.0153\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:44:35 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 400 Bad Request\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 31%|████████████----------------------------| 2287/7340 [78:17<172:59, 29.2 steps/min]2025-08-11 16:44:36,616 - agent.ComputerAgent - INFO - Computer: scroll({'scroll_y': 666, 'scroll_x': 0, 'x': 336, 'y': 152})\n", - "INFO:agent.ComputerAgent:Computer: scroll({'scroll_y': 666, 'scroll_x': 0, 'x': 336, 'y': 152})\n", - "2025-08-11 16:44:37,280 - agent.ComputerAgent - INFO - Computer: click({'x': 520, 'y': 437})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 520, 'y': 437})\n", - "\u001b[92m16:44:37 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m16:44:37 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:44:37,941 - agent.ComputerAgent - INFO - LLM processing started with 42 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 42 messages\n", - "\u001b[92m16:44:37 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 16:44:38,620 - agent.ComputerAgent - INFO - Computer: click({'x': 514, 'y': 304})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 514, 'y': 304})\n", - "2025-08-11 16:44:39,286 - agent.ComputerAgent - INFO - Computer: scroll({'scroll_y': -200, 'scroll_x': 0, 'x': 589, 'y': 128})\n", - "INFO:agent.ComputerAgent:Computer: scroll({'scroll_y': -200, 'scroll_x': 0, 'x': 589, 'y': 128})\n", - " 31%|████████████----------------------------| 2289/7340 [78:21<172:53, 29.2 steps/min]\u001b[92m16:44:39 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:44:39,956 - agent.ComputerAgent - INFO - Computer: click({'x': 351, 'y': 153})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 351, 'y': 153})\n", - "2025-08-11 16:44:40,591 - agent.ComputerAgent - INFO - LLM processing started with 14 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 14 messages\n", - "\u001b[92m16:44:40 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 16:44:41,646 - agent.ComputerAgent - INFO - LLM processing started with 32 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 32 messages\n", - "\u001b[92m16:44:41 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 31%|████████████----------------------------| 2293/7340 [78:23<172:32, 29.3 steps/min]2025-08-11 16:44:42,285 - agent.ComputerAgent - INFO - LLM processing started with 18 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 18 messages\n", - "\u001b[92m16:44:42 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 16:44:42,973 - agent.ComputerAgent - INFO - LLM processing started with 18 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 18 messages\n", - "\u001b[92m16:44:42 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 16:44:44,693 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'enter'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'enter'})\n", - " 31%|████████████----------------------------| 2294/7340 [78:26<172:32, 29.2 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/ed3b76a0-7eeb-473c-8da1-310545515f47/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/fd628f34-1346-4947-bfa4-cf698adb3472/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 16:44:45,371 - agent.ComputerAgent - INFO - LLM processing started with 41 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 41 messages\n", - "\u001b[92m16:44:45 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/5180ec6f-26a5-4ab4-8ca3-87f128083da1/invoke \"HTTP/1.1 200 OK\"\n", - " 31%|████████████----------------------------| 2311/7340 [78:27<170:43, 29.5 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/5180ec6f-26a5-4ab4-8ca3-87f128083da1/close \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/6195bb79-4eff-4d3b-8b67-f28a4e6a73fa/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/fd628f34-1346-4947-bfa4-cf698adb3472/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/2c254802-788e-4b4b-98dc-68cd2c6bcce4/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/74442f45-62aa-40d1-9499-ea3e8e0a4c18/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/d4054e85-5304-43a3-b6d7-128e302780cb/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/7e0b0038-3a97-4d93-8c5c-154cc0b95af9/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/0803e2c2-9de2-40ff-93da-cb49f156cbba/invoke \"HTTP/1.1 200 OK\"\n", - " 31%|████████████----------------------------| 2311/7340 [78:30<170:49, 29.4 steps/min]2025-08-11 16:44:49,145 - agent.ComputerAgent - INFO - LLM processing started with 32 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 32 messages\n", - "\u001b[92m16:44:49 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: GET https://orchestration.hud.so/hud-gym/api/v1/gyms/OSWorld-Ubuntu \"HTTP/1.1 200 OK\"\n", - "2025-08-11 16:44:49,827 - agent.ComputerAgent - INFO - LLM processing started with 18 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 18 messages\n", - "\u001b[92m16:44:49 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 16:44:50,495 - agent.ComputerAgent - INFO - LLM processing started with 22 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 22 messages\n", - "\u001b[92m16:44:50 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/fd628f34-1346-4947-bfa4-cf698adb3472/close \"HTTP/1.1 200 OK\"\n", - " 32%|████████████----------------------------| 2321/7340 [78:32<169:49, 29.6 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/029e3386-8f59-44a0-9dba-b14a4d9825ef/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/a8ddfa4f-6f9b-4ad8-b763-1881394c9926/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/475030e1-7ae3-4ced-91fb-2221d956a2ad/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 400 Bad Request\"\n", - " 32%|████████████----------------------------| 2322/7340 [78:33<169:45, 29.6 steps/min]2025-08-11 16:44:53,297 - agent.ComputerAgent - INFO - LLM processing started with 30 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 30 messages\n", - "\u001b[92m16:44:53 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 16:44:53,963 - agent.ComputerAgent - INFO - LLM processing started with 14 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 14 messages\n", - "\u001b[92m16:44:53 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:44:54 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 32%|████████████----------------------------| 2322/7340 [78:36<169:52, 29.5 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:44:55,334 - agent.ComputerAgent - INFO - LLM processing started with 22 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 22 messages\n", - "\u001b[92m16:44:55 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:accelerate.utils.modeling:We will use 90% of the memory on device 0 for storing the model, and 10% for the buffer to avoid OOM. You can set `max_memory` in to a higher value to use more memory (at your own risk).\n", - "Loading checkpoint shards: 0%| | 0/4 [00:00 ~/vim_test.txt && vim ~/vim_test.txt\\''})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': 'bash -lc \\'printf \"line a\\\\nline b\\\\nline c\\\\n\" > ~/vim_test.txt && vim ~/vim_test.txt\\''})\n", - " 33%|█████████████---------------------------| 2397/7340 [80:43<166:27, 29.7 steps/min]\u001b[92m16:47:01 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:47:02,258 - agent.ComputerAgent - INFO - Computer: click({'x': 855, 'y': 476})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 855, 'y': 476})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:47:02 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 33%|█████████████---------------------------| 2399/7340 [80:44<166:18, 29.7 steps/min]\u001b[92m16:47:03 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:47:03,544 - agent.ComputerAgent - INFO - Computer: scroll({'scroll_y': 604, 'scroll_x': 0, 'x': 307, 'y': 666})\n", - "INFO:agent.ComputerAgent:Computer: scroll({'scroll_y': 604, 'scroll_x': 0, 'x': 307, 'y': 666})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 400 Bad Request\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:47:04 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - " 33%|█████████████---------------------------| 2401/7340 [80:46<166:09, 29.7 steps/min]\u001b[92m16:47:04 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m16:47:04 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:47:05,490 - agent.ComputerAgent - INFO - Agent: Taking a screenshot to see the current computer screen.\n", - "INFO:agent.ComputerAgent:Agent: Taking a screenshot to see the current computer screen.\n", - "2025-08-11 16:47:05,491 - agent.ComputerAgent - INFO - Computer: click({'button': 'left', 'x': 465, 'y': 294})\n", - "INFO:agent.ComputerAgent:Computer: click({'button': 'left', 'x': 465, 'y': 294})\n", - " 33%|█████████████---------------------------| 2402/7340 [80:47<166:05, 29.7 steps/min]\u001b[92m16:47:05 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:47:06,659 - agent.ComputerAgent - INFO - Computer: click({'x': 637, 'y': 471})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 637, 'y': 471})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/3e4ea7d7-21a2-4b07-abd4-a3e280e44e0b/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 16:47:07,327 - agent.ComputerAgent - INFO - LLM processing started with 31 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 31 messages\n", - "\u001b[92m16:47:07 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 33%|█████████████---------------------------| 2403/7340 [80:49<166:02, 29.7 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m16:47:07 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:47:08,385 - agent.ComputerAgent - INFO - Computer: click({'x': 111, 'y': 270})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 111, 'y': 270})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/6195bb79-4eff-4d3b-8b67-f28a4e6a73fa/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/475030e1-7ae3-4ced-91fb-2221d956a2ad/invoke \"HTTP/1.1 200 OK\"\n", - " 33%|█████████████---------------------------| 2404/7340 [80:50<165:58, 29.7 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:47:09,003 - agent.ComputerAgent - INFO - LLM processing started with 30 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 30 messages\n", - "\u001b[92m16:47:09 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/6fcb07bb-6857-4888-82a0-1fd0dbf2d722/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m16:47:09 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:47:09,698 - agent.ComputerAgent - INFO - Computer: scroll({'scroll_y': 530, 'scroll_x': 0, 'x': 574, 'y': 736})\n", - "INFO:agent.ComputerAgent:Computer: scroll({'scroll_y': 530, 'scroll_x': 0, 'x': 574, 'y': 736})\n", - "2025-08-11 16:47:10,366 - agent.ComputerAgent - INFO - LLM processing started with 32 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 32 messages\n", - "\u001b[92m16:47:10 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 400 Bad Request\"\n", - " 33%|█████████████---------------------------| 2406/7340 [80:52<165:50, 29.8 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/create_environment \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:47:10 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:47:11,392 - agent.ComputerAgent - INFO - Computer: click({'x': 1008, 'y': 164})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 1008, 'y': 164})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/d4054e85-5304-43a3-b6d7-128e302780cb/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/81b23870-39ed-4649-9729-1d4809f713ec/invoke \"HTTP/1.1 200 OK\"\n", - " 33%|█████████████---------------------------| 2407/7340 [80:53<165:46, 29.8 steps/min]2025-08-11 16:47:12,011 - agent.ComputerAgent - INFO - LLM processing started with 6 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 6 messages\n", - "\u001b[92m16:47:12 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 16:47:12,685 - agent.ComputerAgent - INFO - LLM processing started with 42 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 42 messages\n", - "\u001b[92m16:47:12 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/3e4ea7d7-21a2-4b07-abd4-a3e280e44e0b/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:47:13 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 33%|█████████████---------------------------| 2408/7340 [80:55<165:44, 29.8 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:47:14,028 - agent.ComputerAgent - INFO - LLM processing started with 33 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 33 messages\n", - "\u001b[92m16:47:14 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/7e0b0038-3a97-4d93-8c5c-154cc0b95af9/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/1473c3f2-39e1-4aff-8d55-0e23dc25a055/invoke \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:47:14,693 - agent.ComputerAgent - INFO - LLM processing started with 6 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 6 messages\n", - "\u001b[92m16:47:14 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - " 33%|█████████████---------------------------| 2408/7340 [80:57<165:48, 29.7 steps/min]\u001b[92m16:47:15 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m16:47:15 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/a46ee6f6-d167-47c4-ad83-e16b88450253/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:47:16,409 - agent.ComputerAgent - INFO - Computer: click({'x': 237, 'y': 75})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 237, 'y': 75})\n", - "2025-08-11 16:47:17,070 - agent.ComputerAgent - INFO - LLM processing started with 40 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 40 messages\n", - "\u001b[92m16:47:17 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 16:47:18,388 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'ctrl+a'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'ctrl+a'})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 400 Bad Request\"\n", - "\u001b[92m16:47:18 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/029e3386-8f59-44a0-9dba-b14a4d9825ef/invoke \"HTTP/1.1 200 OK\"\n", - " 33%|█████████████---------------------------| 2409/7340 [81:00<165:48, 29.7 steps/min]2025-08-11 16:47:19,055 - agent.ComputerAgent - INFO - Computer: click({'x': 458, 'y': 275})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 458, 'y': 275})\n", - "2025-08-11 16:47:19,739 - agent.ComputerAgent - INFO - LLM processing started with 26 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 26 messages\n", - "\u001b[92m16:47:19 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 33%|█████████████---------------------------| 2410/7340 [81:01<165:44, 29.7 steps/min]2025-08-11 16:47:20,387 - agent.ComputerAgent - INFO - LLM processing started with 36 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 36 messages\n", - "\u001b[92m16:47:20 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 16:47:21,068 - agent.ComputerAgent - INFO - LLM processing started with 24 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 24 messages\n", - "\u001b[92m16:47:21 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 33%|█████████████---------------------------| 2411/7340 [81:02<165:41, 29.7 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:47:21 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/ff8b808f-c3a6-4979-8f9a-c6a25905116c/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/3e4ea7d7-21a2-4b07-abd4-a3e280e44e0b/invoke \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:47:22,362 - agent.ComputerAgent - INFO - LLM processing started with 35 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 35 messages\n", - "\u001b[92m16:47:22 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 16:47:23,717 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'enter'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'enter'})\n", - " 33%|█████████████---------------------------| 2411/7340 [81:05<165:46, 29.7 steps/min]\u001b[92m16:47:23 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:47:24,730 - agent.ComputerAgent - INFO - Computer: double_click({'x': 331, 'y': 111})\n", - "INFO:agent.ComputerAgent:Computer: double_click({'x': 331, 'y': 111})\n", - " 33%|█████████████---------------------------| 2412/7340 [81:06<165:42, 29.7 steps/min]2025-08-11 16:47:25,410 - agent.ComputerAgent - INFO - LLM processing started with 38 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 38 messages\n", - "\u001b[92m16:47:25 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v1/environments/e2ffab0a-c998-4bbf-906b-d3aad0586220/reset \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/2c254802-788e-4b4b-98dc-68cd2c6bcce4/invoke \"HTTP/1.1 200 OK\"\n", - " 33%|█████████████---------------------------| 2413/7340 [81:07<165:38, 29.7 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/0803e2c2-9de2-40ff-93da-cb49f156cbba/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 16:47:26,610 - agent.ComputerAgent - INFO - LLM processing started with 26 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 26 messages\n", - "\u001b[92m16:47:26 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 16:47:27,269 - agent.ComputerAgent - INFO - LLM processing started with 40 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 40 messages\n", - "\u001b[92m16:47:27 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 33%|█████████████---------------------------| 2413/7340 [81:09<165:43, 29.7 steps/min]\u001b[92m16:47:27 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/e2ffab0a-c998-4bbf-906b-d3aad0586220/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 400 Bad Request\"\n", - "\u001b[92m16:47:28 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:47:29,307 - agent.ComputerAgent - INFO - LLM processing started with 1 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 1 messages\n", - "\u001b[92m16:47:29 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 33%|█████████████---------------------------| 2414/7340 [81:11<165:39, 29.7 steps/min]\u001b[92m16:47:30 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:47:30,494 - agent.ComputerAgent - INFO - Computer: click({'x': 946, 'y': 738})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 946, 'y': 738})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/475030e1-7ae3-4ced-91fb-2221d956a2ad/invoke \"HTTP/1.1 200 OK\"\n", - " 33%|█████████████---------------------------| 2414/7340 [81:12<165:42, 29.7 steps/min]2025-08-11 16:47:31,166 - agent.ComputerAgent - INFO - LLM processing started with 34 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 34 messages\n", - "\u001b[92m16:47:31 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/3e4ea7d7-21a2-4b07-abd4-a3e280e44e0b/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 16:47:31,814 - agent.ComputerAgent - INFO - LLM processing started with 37 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 37 messages\n", - "\u001b[92m16:47:31 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m16:47:31 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/6fcb07bb-6857-4888-82a0-1fd0dbf2d722/invoke \"HTTP/1.1 200 OK\"\n", - " 33%|█████████████---------------------------| 2415/7340 [81:13<165:39, 29.7 steps/min]2025-08-11 16:47:32,487 - agent.ComputerAgent - INFO - Computer: click({'x': 351, 'y': 294})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 351, 'y': 294})\n", - "2025-08-11 16:47:33,177 - agent.ComputerAgent - INFO - LLM processing started with 8 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 8 messages\n", - "\u001b[92m16:47:33 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 33%|█████████████---------------------------| 2415/7340 [81:14<165:41, 29.7 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 33%|█████████████---------------------------| 2416/7340 [81:15<165:37, 29.7 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/create_environment \"HTTP/1.1 200 OK\"\n", - " 33%|█████████████---------------------------| 2416/7340 [81:16<165:39, 29.7 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/e5b48f23-7687-494c-b68c-ebdfc70d085f/invoke \"HTTP/1.1 200 OK\"\n", - " 33%|█████████████---------------------------| 2416/7340 [81:17<165:41, 29.7 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 400 Bad Request\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/create_environment \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:47:37 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 33%|█████████████---------------------------| 2417/7340 [81:19<165:37, 29.7 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/3e4ea7d7-21a2-4b07-abd4-a3e280e44e0b/invoke \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:47:38,035 - agent.ComputerAgent - INFO - LLM processing started with 39 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 39 messages\n", - "\u001b[92m16:47:38 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/6195bb79-4eff-4d3b-8b67-f28a4e6a73fa/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v1/environments/e5b48f23-7687-494c-b68c-ebdfc70d085f/reset \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/c3518cd0-0df6-44e9-8393-0c62002bc984/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/81b23870-39ed-4649-9729-1d4809f713ec/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 16:47:38,680 - agent.ComputerAgent - INFO - LLM processing started with 32 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 32 messages\n", - "\u001b[92m16:47:38 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 33%|█████████████---------------------------| 2417/7340 [81:20<165:40, 29.7 steps/min]2025-08-11 16:47:39,336 - agent.ComputerAgent - INFO - LLM processing started with 8 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 8 messages\n", - "\u001b[92m16:47:39 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m16:47:39 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m16:47:40 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:47:41,078 - agent.ComputerAgent - INFO - Agent: Taking a screenshot to see the current computer screen.\n", - "INFO:agent.ComputerAgent:Agent: Taking a screenshot to see the current computer screen.\n", - "2025-08-11 16:47:41,079 - agent.ComputerAgent - INFO - Computer: click({'x': 92, 'y': 359})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 92, 'y': 359})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - " 33%|█████████████---------------------------| 2417/7340 [81:23<165:46, 29.7 steps/min]\u001b[92m16:47:41 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - " 33%|█████████████---------------------------| 2418/7340 [81:24<165:42, 29.7 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m16:47:43 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m16:47:43 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:47:44,087 - agent.ComputerAgent - INFO - Computer: click({'x': 982, 'y': 760})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 982, 'y': 760})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/e5b48f23-7687-494c-b68c-ebdfc70d085f/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 400 Bad Request\"\n", - " 33%|█████████████---------------------------| 2419/7340 [81:25<165:39, 29.7 steps/min]2025-08-11 16:47:44,730 - agent.ComputerAgent - INFO - LLM processing started with 1 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 1 messages\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m16:47:44 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m16:47:44 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m16:47:45 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:47:46,038 - agent.ComputerAgent - INFO - Computer: scroll({'scroll_y': 654, 'scroll_x': 0, 'x': 283, 'y': 664})\n", - "INFO:agent.ComputerAgent:Computer: scroll({'scroll_y': 654, 'scroll_x': 0, 'x': 283, 'y': 664})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - " 33%|█████████████---------------------------| 2420/7340 [81:28<165:38, 29.7 steps/min]\u001b[92m16:47:46 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m16:47:46 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/e2ffab0a-c998-4bbf-906b-d3aad0586220/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:47:47 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:47:47,983 - agent.ComputerAgent - INFO - Computer: click({'x': 585, 'y': 355})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 585, 'y': 355})\n", - "\u001b[92m16:47:47 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 33%|█████████████---------------------------| 2421/7340 [81:29<165:34, 29.7 steps/min]2025-08-11 16:47:48,672 - agent.ComputerAgent - INFO - Computer: click({'x': 962, 'y': 234})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 962, 'y': 234})\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:47:49,341 - agent.ComputerAgent - INFO - LLM processing started with 6 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 6 messages\n", - "\u001b[92m16:47:49 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m16:47:49 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 33%|█████████████---------------------------| 2422/7340 [81:31<165:31, 29.7 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:47:50,011 - agent.ComputerAgent - INFO - Computer: click({'x': 392, 'y': 275})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 392, 'y': 275})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 16:47:51,361 - agent.ComputerAgent - INFO - Agent: Taking a screenshot to see the current computer screen.\n", - "INFO:agent.ComputerAgent:Agent: Taking a screenshot to see the current computer screen.\n", - "2025-08-11 16:47:51,362 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'super'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'super'})\n", - " 33%|█████████████---------------------------| 2423/7340 [81:33<165:29, 29.7 steps/min]\u001b[92m16:47:51 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/3e4ea7d7-21a2-4b07-abd4-a3e280e44e0b/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:47:52,058 - agent.ComputerAgent - INFO - LLM processing started with 41 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 41 messages\n", - "\u001b[92m16:47:52 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 16:47:52,742 - agent.ComputerAgent - INFO - Computer: click({'x': 196, 'y': 237})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 196, 'y': 237})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:47:53 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 33%|█████████████---------------------------| 2425/7340 [81:35<165:21, 29.7 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:47:54 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 33%|█████████████---------------------------| 2426/7340 [81:36<165:17, 29.7 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/7e0b0038-3a97-4d93-8c5c-154cc0b95af9/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/029e3386-8f59-44a0-9dba-b14a4d9825ef/invoke \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:47:54 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:47:55,312 - agent.ComputerAgent - INFO - Computer: click({'x': 351, 'y': 294})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 351, 'y': 294})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/a39ee9df-d3ba-456a-95cf-3a11a826583b/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 400 Bad Request\"\n", - "2025-08-11 16:47:55,965 - agent.ComputerAgent - INFO - LLM processing started with 42 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 42 messages\n", - "\u001b[92m16:47:56 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 16:47:57,620 - agent.ComputerAgent - INFO - Computer: wait({})\n", - "INFO:agent.ComputerAgent:Computer: wait({})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/d4054e85-5304-43a3-b6d7-128e302780cb/invoke \"HTTP/1.1 200 OK\"\n", - " 33%|█████████████---------------------------| 2427/7340 [81:39<165:17, 29.7 steps/min]\u001b[92m16:47:57 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:47:58,279 - agent.ComputerAgent - INFO - LLM processing started with 38 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 38 messages\n", - "\u001b[92m16:47:58 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 16:47:58,928 - agent.ComputerAgent - INFO - Computer: click({'x': 316, 'y': 101})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 316, 'y': 101})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/0803e2c2-9de2-40ff-93da-cb49f156cbba/invoke \"HTTP/1.1 200 OK\"\n", - " 33%|█████████████---------------------------| 2429/7340 [81:40<165:08, 29.7 steps/min]2025-08-11 16:47:59,585 - agent.ComputerAgent - INFO - LLM processing started with 28 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 28 messages\n", - "\u001b[92m16:47:59 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 16:48:00,942 - agent.ComputerAgent - INFO - Computer: type({'text': ':q'})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': ':q'})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:48:01 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:48:02,299 - agent.ComputerAgent - INFO - LLM processing started with 28 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 28 messages\n", - "\u001b[92m16:48:02 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/6fcb07bb-6857-4888-82a0-1fd0dbf2d722/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/e5b48f23-7687-494c-b68c-ebdfc70d085f/invoke \"HTTP/1.1 200 OK\"\n", - " 33%|█████████████---------------------------| 2430/7340 [81:44<165:09, 29.7 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m16:48:03 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 33%|█████████████---------------------------| 2431/7340 [81:45<165:04, 29.7 steps/min]2025-08-11 16:48:03,692 - agent.ComputerAgent - INFO - LLM processing started with 10 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 10 messages\n", - "\u001b[92m16:48:03 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/3e4ea7d7-21a2-4b07-abd4-a3e280e44e0b/invoke \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:48:04,331 - agent.ComputerAgent - INFO - LLM processing started with 43 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 43 messages\n", - "\u001b[92m16:48:04 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 33%|█████████████---------------------------| 2431/7340 [81:46<165:07, 29.7 steps/min]\u001b[92m16:48:04 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:48:04,993 - agent.ComputerAgent - INFO - Computer: click({'x': 458, 'y': 422})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 458, 'y': 422})\n", - "2025-08-11 16:48:05,658 - agent.ComputerAgent - INFO - LLM processing started with 6 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 6 messages\n", - "\u001b[92m16:48:05 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/d4054e85-5304-43a3-b6d7-128e302780cb/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m16:48:05 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 33%|█████████████---------------------------| 2431/7340 [81:47<165:09, 29.7 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:48:07,064 - agent.ComputerAgent - INFO - Computer: click({'x': 474, 'y': 332})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 474, 'y': 332})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/81b23870-39ed-4649-9729-1d4809f713ec/invoke \"HTTP/1.1 200 OK\"\n", - " 33%|█████████████---------------------------| 2436/7340 [81:48<164:42, 29.8 steps/min]2025-08-11 16:48:07,712 - agent.ComputerAgent - INFO - LLM processing started with 10 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 10 messages\n", - "\u001b[92m16:48:07 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/475030e1-7ae3-4ced-91fb-2221d956a2ad/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 16:48:08,382 - agent.ComputerAgent - INFO - LLM processing started with 36 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 36 messages\n", - "\u001b[92m16:48:08 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/a46ee6f6-d167-47c4-ad83-e16b88450253/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/d4054e85-5304-43a3-b6d7-128e302780cb/close \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 400 Bad Request\"\n", - " 33%|█████████████---------------------------| 2438/7340 [81:50<164:32, 29.8 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/6195bb79-4eff-4d3b-8b67-f28a4e6a73fa/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 16:48:09,668 - agent.ComputerAgent - INFO - LLM processing started with 26 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 26 messages\n", - "\u001b[92m16:48:09 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: GET https://orchestration.hud.so/hud-gym/api/v1/gyms/OSWorld-Ubuntu \"HTTP/1.1 200 OK\"\n", - " 33%|█████████████---------------------------| 2438/7340 [81:51<164:35, 29.8 steps/min]2025-08-11 16:48:10,320 - agent.ComputerAgent - INFO - LLM processing started with 34 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 34 messages\n", - "\u001b[92m16:48:10 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/3e4ea7d7-21a2-4b07-abd4-a3e280e44e0b/invoke \"HTTP/1.1 200 OK\"\n", - " 33%|█████████████---------------------------| 2438/7340 [81:52<164:37, 29.8 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v1/environments/1473c3f2-39e1-4aff-8d55-0e23dc25a055/reset \"HTTP/1.1 200 OK\"\n", - " 33%|█████████████---------------------------| 2438/7340 [81:53<164:39, 29.8 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/ff8b808f-c3a6-4979-8f9a-c6a25905116c/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/e2ffab0a-c998-4bbf-906b-d3aad0586220/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 16:48:12,541 - agent.ComputerAgent - INFO - LLM processing started with 40 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 40 messages\n", - "\u001b[92m16:48:12 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 16:48:13,223 - agent.ComputerAgent - INFO - LLM processing started with 8 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 8 messages\n", - "\u001b[92m16:48:13 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 33%|█████████████---------------------------| 2438/7340 [81:54<164:42, 29.8 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/1473c3f2-39e1-4aff-8d55-0e23dc25a055/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 16:48:14,391 - agent.ComputerAgent - INFO - LLM processing started with 1 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 1 messages\n", - "\u001b[92m16:48:14 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 33%|█████████████---------------------------| 2438/7340 [81:56<164:44, 29.8 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/3e4ea7d7-21a2-4b07-abd4-a3e280e44e0b/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:48:15 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 33%|█████████████---------------------------| 2440/7340 [81:57<164:35, 29.8 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/3e4ea7d7-21a2-4b07-abd4-a3e280e44e0b/close \"HTTP/1.1 200 OK\"\n", - "Loading checkpoint shards: 0%| | 0/4 [00:005'})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': 'WEEKDAY(B3;2)>5'})\n", - " 34%|█████████████---------------------------| 2477/7340 [83:44<164:25, 29.6 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:50:03 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/6195bb79-4eff-4d3b-8b67-f28a4e6a73fa/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v1/environments/422a20c8-b318-46e4-9f06-d599c9ed261c/reset \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:50:04 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m16:50:04 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 34%|█████████████---------------------------| 2479/7340 [83:46<164:15, 29.6 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:50:05,091 - agent.ComputerAgent - INFO - Computer: click({'x': 351, 'y': 294})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 351, 'y': 294})\n", - "\u001b[92m16:50:05 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m16:50:05 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:50:05,754 - agent.ComputerAgent - INFO - LLM processing started with 40 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 40 messages\n", - "\u001b[92m16:50:05 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - " 34%|█████████████---------------------------| 2479/7340 [83:48<164:19, 29.6 steps/min]\u001b[92m16:50:06 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m16:50:06 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:50:07,118 - agent.ComputerAgent - INFO - Computer: drag({'path': [{'x': 471, 'y': 328}, {'x': 351, 'y': 709}]})\n", - "INFO:agent.ComputerAgent:Computer: drag({'path': [{'x': 471, 'y': 328}, {'x': 351, 'y': 709}]})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:50:07 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m16:50:07 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/422a20c8-b318-46e4-9f06-d599c9ed261c/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:50:08,416 - agent.ComputerAgent - INFO - Computer: click({'x': 268, 'y': 188})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 268, 'y': 188})\n", - "2025-08-11 16:50:09,045 - agent.ComputerAgent - INFO - LLM processing started with 1 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 1 messages\n", - "\u001b[92m16:50:09 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 34%|█████████████---------------------------| 2480/7340 [83:50<164:18, 29.6 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:50:09,730 - agent.ComputerAgent - INFO - LLM processing started with 8 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 8 messages\n", - "\u001b[92m16:50:09 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m16:50:09 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/0803e2c2-9de2-40ff-93da-cb49f156cbba/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:50:10,446 - agent.ComputerAgent - INFO - Computer: double_click({'x': 618, 'y': 483})\n", - "INFO:agent.ComputerAgent:Computer: double_click({'x': 618, 'y': 483})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/e5b48f23-7687-494c-b68c-ebdfc70d085f/invoke \"HTTP/1.1 200 OK\"\n", - " 34%|█████████████---------------------------| 2482/7340 [83:52<164:09, 29.6 steps/min]2025-08-11 16:50:11,451 - agent.ComputerAgent - INFO - LLM processing started with 34 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 34 messages\n", - "\u001b[92m16:50:11 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/f73836c4-d8e3-425b-a750-f2319c89164e/invoke \"HTTP/1.1 200 OK\"\n", - " 34%|█████████████---------------------------| 2483/7340 [83:53<164:05, 29.6 steps/min]2025-08-11 16:50:12,123 - agent.ComputerAgent - INFO - LLM processing started with 14 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 14 messages\n", - "\u001b[92m16:50:12 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 16:50:13,563 - agent.ComputerAgent - INFO - LLM processing started with 12 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 12 messages\n", - "\u001b[92m16:50:13 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 34%|█████████████---------------------------| 2483/7340 [83:55<164:09, 29.6 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v1/environments/514e0362-c0b3-4216-989f-d260ec405efb/reset \"HTTP/1.1 200 OK\"\n", - " 34%|█████████████---------------------------| 2483/7340 [83:56<164:11, 29.6 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/81b23870-39ed-4649-9729-1d4809f713ec/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 16:50:15,740 - agent.ComputerAgent - INFO - LLM processing started with 18 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 18 messages\n", - "\u001b[92m16:50:15 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 34%|█████████████---------------------------| 2483/7340 [83:57<164:13, 29.6 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 16:50:17,970 - agent.ComputerAgent - INFO - Agent: Taking a screenshot to see the current computer screen.\n", - "INFO:agent.ComputerAgent:Agent: Taking a screenshot to see the current computer screen.\n", - "2025-08-11 16:50:17,970 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'win+e'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'win+e'})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/514e0362-c0b3-4216-989f-d260ec405efb/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/a39ee9df-d3ba-456a-95cf-3a11a826583b/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/1473c3f2-39e1-4aff-8d55-0e23dc25a055/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/e2ffab0a-c998-4bbf-906b-d3aad0586220/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - " 34%|█████████████---------------------------| 2483/7340 [84:00<164:19, 29.6 steps/min]\u001b[92m16:50:18 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 16:50:19,965 - agent.ComputerAgent - INFO - Computer: type({'text': 'Orchis theme gnome-look'})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': 'Orchis theme gnome-look'})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:50:20 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:50:21,285 - agent.ComputerAgent - INFO - LLM processing started with 6 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 6 messages\n", - "\u001b[92m16:50:21 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m16:50:21 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 34%|█████████████---------------------------| 2483/7340 [84:03<164:25, 29.5 steps/min]\u001b[92m16:50:22 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:50:22,607 - agent.ComputerAgent - INFO - LLM processing started with 18 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 18 messages\n", - "\u001b[92m16:50:22 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 16:50:23,283 - agent.ComputerAgent - INFO - Computer: scroll({'scroll_y': 626, 'scroll_x': 0, 'x': 588, 'y': 446})\n", - "INFO:agent.ComputerAgent:Computer: scroll({'scroll_y': 626, 'scroll_x': 0, 'x': 588, 'y': 446})\n", - "\u001b[92m16:50:23 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v1/environments/daac505f-9423-4b29-b11c-9b23c5c9e3ee/reset \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:50:23 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:50:23,909 - agent.ComputerAgent - INFO - Agent: Taking a screenshot to see the current computer screen.\n", - "INFO:agent.ComputerAgent:Agent: Taking a screenshot to see the current computer screen.\n", - "2025-08-11 16:50:23,909 - agent.ComputerAgent - INFO - Computer: double_click({'x': 989, 'y': 713})\n", - "INFO:agent.ComputerAgent:Computer: double_click({'x': 989, 'y': 713})\n", - " 34%|█████████████---------------------------| 2484/7340 [84:05<164:23, 29.5 steps/min]2025-08-11 16:50:24,578 - agent.ComputerAgent - INFO - Computer: click({'x': 412, 'y': 128})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 412, 'y': 128})\n", - "2025-08-11 16:50:25,256 - agent.ComputerAgent - INFO - LLM processing started with 10 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 10 messages\n", - "\u001b[92m16:50:25 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 34%|█████████████---------------------------| 2486/7340 [84:07<164:14, 29.6 steps/min]2025-08-11 16:50:25,924 - agent.ComputerAgent - INFO - LLM processing started with 34 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 34 messages\n", - "\u001b[92m16:50:25 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 16:50:26,594 - agent.ComputerAgent - INFO - LLM processing started with 1 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 1 messages\n", - "\u001b[92m16:50:26 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - " 34%|█████████████---------------------------| 2487/7340 [84:09<164:12, 29.6 steps/min]\u001b[92m16:50:27 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m16:50:27 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:50:28,422 - agent.ComputerAgent - INFO - Computer: double_click({'x': 960, 'y': 713})\n", - "INFO:agent.ComputerAgent:Computer: double_click({'x': 960, 'y': 713})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 16:50:29,722 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'win'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'win'})\n", - " 34%|█████████████---------------------------| 2487/7340 [84:11<164:17, 29.5 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/daac505f-9423-4b29-b11c-9b23c5c9e3ee/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 16:50:30,365 - agent.ComputerAgent - INFO - LLM processing started with 1 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 1 messages\n", - "\u001b[92m16:50:30 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/fed9747f-6005-4d29-b83e-afc7934c0ff5/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:50:31 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/e5b48f23-7687-494c-b68c-ebdfc70d085f/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - " 34%|█████████████---------------------------| 2489/7340 [84:13<164:09, 29.6 steps/min]\u001b[92m16:50:31 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/2b43eb21-4025-495a-8c66-358bfcac034b/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m16:50:32 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/029e3386-8f59-44a0-9dba-b14a4d9825ef/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 16:50:33,742 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'esc'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'esc'})\n", - "\u001b[92m16:50:33 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m16:50:33 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - " 34%|█████████████---------------------------| 2489/7340 [84:16<164:14, 29.5 steps/min]\u001b[92m16:50:34 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:50:35,048 - agent.ComputerAgent - INFO - LLM processing started with 6 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 6 messages\n", - "\u001b[92m16:50:35 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 16:50:35,696 - agent.ComputerAgent - INFO - Computer: move({'x': 887, 'y': 167})\n", - "INFO:agent.ComputerAgent:Computer: move({'x': 887, 'y': 167})\n", - "2025-08-11 16:50:36,379 - agent.ComputerAgent - INFO - Computer: click({'x': 260, 'y': 101})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 260, 'y': 101})\n", - "\u001b[92m16:50:36 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:50:37,041 - agent.ComputerAgent - INFO - Computer: click({'x': 537, 'y': 304})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 537, 'y': 304})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 16:50:38,373 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'right'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'right'})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/f73836c4-d8e3-425b-a750-f2319c89164e/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/422a20c8-b318-46e4-9f06-d599c9ed261c/invoke \"HTTP/1.1 200 OK\"\n", - " 34%|█████████████---------------------------| 2490/7340 [84:20<164:16, 29.5 steps/min]\u001b[92m16:50:38 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:50:39,696 - agent.ComputerAgent - INFO - Computer: type({'text': 'Mumbai'})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': 'Mumbai'})\n", - "2025-08-11 16:50:40,392 - agent.ComputerAgent - INFO - Computer: click({'x': 746, 'y': 651})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 746, 'y': 651})\n", - " 34%|█████████████---------------------------| 2494/7340 [84:22<163:56, 29.6 steps/min]2025-08-11 16:50:41,039 - agent.ComputerAgent - INFO - LLM processing started with 8 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 8 messages\n", - "\u001b[92m16:50:41 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/a46ee6f6-d167-47c4-ad83-e16b88450253/invoke \"HTTP/1.1 502 Bad Gateway\"\n", - "2025-08-11 16:50:41,701 - agent.ComputerAgent - INFO - LLM processing started with 10 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 10 messages\n", - "\u001b[92m16:50:41 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 16:50:42,355 - agent.ComputerAgent - INFO - LLM processing started with 16 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 16 messages\n", - "\u001b[92m16:50:42 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 34%|█████████████---------------------------| 2496/7340 [84:24<163:47, 29.6 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 16:50:43,043 - agent.ComputerAgent - INFO - LLM processing started with 14 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 14 messages\n", - "\u001b[92m16:50:43 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 34%|█████████████---------------------------| 2496/7340 [84:25<163:49, 29.6 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 16:50:45,382 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'ctrl+home'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'ctrl+home'})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/0803e2c2-9de2-40ff-93da-cb49f156cbba/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/1473c3f2-39e1-4aff-8d55-0e23dc25a055/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/ff8b808f-c3a6-4979-8f9a-c6a25905116c/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/81b23870-39ed-4649-9729-1d4809f713ec/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/e2ffab0a-c998-4bbf-906b-d3aad0586220/invoke \"HTTP/1.1 200 OK\"\n", - " 34%|█████████████---------------------------| 2496/7340 [84:27<163:53, 29.6 steps/min]2025-08-11 16:50:46,040 - agent.ComputerAgent - INFO - LLM processing started with 20 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 20 messages\n", - "\u001b[92m16:50:46 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 16:50:46,708 - agent.ComputerAgent - INFO - LLM processing started with 20 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 20 messages\n", - "\u001b[92m16:50:46 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 16:50:47,380 - agent.ComputerAgent - INFO - LLM processing started with 8 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 8 messages\n", - "\u001b[92m16:50:47 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 34%|█████████████---------------------------| 2496/7340 [84:29<163:57, 29.5 steps/min]2025-08-11 16:50:48,016 - agent.ComputerAgent - INFO - LLM processing started with 36 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 36 messages\n", - "\u001b[92m16:50:48 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 16:50:49,339 - agent.ComputerAgent - INFO - Agent: Taking a screenshot to see the current computer screen.\n", - "INFO:agent.ComputerAgent:Agent: Taking a screenshot to see the current computer screen.\n", - "2025-08-11 16:50:49,340 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'ctrl+f'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'ctrl+f'})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/a46ee6f6-d167-47c4-ad83-e16b88450253/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/6195bb79-4eff-4d3b-8b67-f28a4e6a73fa/invoke \"HTTP/1.1 200 OK\"\n", - " 34%|█████████████---------------------------| 2496/7340 [84:31<164:01, 29.5 steps/min]2025-08-11 16:50:49,998 - agent.ComputerAgent - INFO - LLM processing started with 12 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 12 messages\n", - "\u001b[92m16:50:50 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:50:50 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:50:51,356 - agent.ComputerAgent - INFO - LLM processing started with 6 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 6 messages\n", - "\u001b[92m16:50:51 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:50:52 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 34%|█████████████---------------------------| 2496/7340 [84:34<164:08, 29.5 steps/min]\u001b[92m16:50:52 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:50:52 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m16:50:53 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:50:54,063 - agent.ComputerAgent - INFO - Computer: click({'button': 'left', 'x': 398, 'y': 89})\n", - "INFO:agent.ComputerAgent:Computer: click({'button': 'left', 'x': 398, 'y': 89})\n", - "\u001b[92m16:50:54 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m16:50:54 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:50:54,739 - agent.ComputerAgent - INFO - Computer: double_click({'x': 960, 'y': 713})\n", - "INFO:agent.ComputerAgent:Computer: double_click({'x': 960, 'y': 713})\n", - " 34%|█████████████---------------------------| 2496/7340 [84:36<164:11, 29.5 steps/min]2025-08-11 16:50:55,391 - agent.ComputerAgent - INFO - Computer: click({'x': 793, 'y': 41})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 793, 'y': 41})\n", - "\u001b[92m16:50:55 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:50:56,056 - agent.ComputerAgent - INFO - LLM processing started with 42 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 42 messages\n", - "\u001b[92m16:50:56 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 16:50:56,694 - agent.ComputerAgent - INFO - Computer: click({'x': 17, 'y': 427})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 17, 'y': 427})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:50:57 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 34%|█████████████---------------------------| 2498/7340 [84:39<164:06, 29.5 steps/min]\u001b[92m16:50:58 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:50:58,732 - agent.ComputerAgent - INFO - LLM processing started with 32 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 32 messages\n", - "\u001b[92m16:50:58 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m16:50:58 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:50:59,814 - agent.ComputerAgent - INFO - Computer: click({'x': 318, 'y': 237})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 318, 'y': 237})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:50:59 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m16:51:00 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:51:02,000 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'right'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'right'})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/029e3386-8f59-44a0-9dba-b14a4d9825ef/invoke \"HTTP/1.1 200 OK\"\n", - " 34%|█████████████---------------------------| 2500/7340 [84:43<164:02, 29.5 steps/min]2025-08-11 16:51:02,679 - agent.ComputerAgent - INFO - Agent: Taking a screenshot to see the current computer screen.\n", - "INFO:agent.ComputerAgent:Agent: Taking a screenshot to see the current computer screen.\n", - "2025-08-11 16:51:02,680 - agent.ComputerAgent - INFO - Computer: click({'x': 95, 'y': 185})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 95, 'y': 185})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:51:04,082 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'ctrl+home'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'ctrl+home'})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:51:04 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m16:51:04 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 34%|█████████████---------------------------| 2502/7340 [84:46<163:55, 29.5 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:51:05,471 - agent.ComputerAgent - INFO - LLM processing started with 10 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 10 messages\n", - "\u001b[92m16:51:05 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m16:51:05 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m16:51:05 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 16:51:06,153 - agent.ComputerAgent - INFO - Computer: double_click({'x': 615, 'y': 483})\n", - "INFO:agent.ComputerAgent:Computer: double_click({'x': 615, 'y': 483})\n", - "\u001b[92m16:51:06 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 34%|█████████████---------------------------| 2503/7340 [84:47<163:52, 29.5 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 16:51:06,824 - agent.ComputerAgent - INFO - Computer: drag({'path': [{'x': 483, 'y': 328}, {'x': 411, 'y': 711}]})\n", - "INFO:agent.ComputerAgent:Computer: drag({'path': [{'x': 483, 'y': 328}, {'x': 411, 'y': 711}]})\n", - " 34%|█████████████---------------------------| 2504/7340 [84:48<163:48, 29.5 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/029e3386-8f59-44a0-9dba-b14a4d9825ef/close \"HTTP/1.1 200 OK\"\n", - " 34%|█████████████---------------------------| 2505/7340 [84:49<163:44, 29.5 steps/min]INFO:httpx:HTTP Request: GET https://orchestration.hud.so/hud-gym/api/v1/gyms/OSWorld-Ubuntu \"HTTP/1.1 200 OK\"\n", - " 34%|█████████████---------------------------| 2505/7340 [84:50<163:46, 29.5 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/e5b48f23-7687-494c-b68c-ebdfc70d085f/invoke \"HTTP/1.1 200 OK\"\n", - " 34%|█████████████---------------------------| 2505/7340 [84:51<163:48, 29.5 steps/min]2025-08-11 16:51:11,251 - agent.ComputerAgent - INFO - LLM processing started with 18 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 18 messages\n", - "\u001b[92m16:51:11 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/2b43eb21-4025-495a-8c66-358bfcac034b/invoke \"HTTP/1.1 200 OK\"\n", - " 34%|█████████████---------------------------| 2505/7340 [84:53<163:50, 29.5 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/daac505f-9423-4b29-b11c-9b23c5c9e3ee/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/f73836c4-d8e3-425b-a750-f2319c89164e/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/81b23870-39ed-4649-9729-1d4809f713ec/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/422a20c8-b318-46e4-9f06-d599c9ed261c/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/1473c3f2-39e1-4aff-8d55-0e23dc25a055/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 16:51:11,946 - agent.ComputerAgent - INFO - LLM processing started with 6 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 6 messages\n", - "\u001b[92m16:51:11 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 16:51:12,584 - agent.ComputerAgent - INFO - LLM processing started with 22 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 22 messages\n", - "\u001b[92m16:51:12 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/e2ffab0a-c998-4bbf-906b-d3aad0586220/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/a39ee9df-d3ba-456a-95cf-3a11a826583b/invoke \"HTTP/1.1 200 OK\"\n", - " 34%|█████████████---------------------------| 2505/7340 [84:54<163:52, 29.5 steps/min]2025-08-11 16:51:13,334 - agent.ComputerAgent - INFO - LLM processing started with 12 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 12 messages\n", - "\u001b[92m16:51:13 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 16:51:13,968 - agent.ComputerAgent - INFO - LLM processing started with 22 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 22 messages\n", - "\u001b[92m16:51:13 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 34%|█████████████---------------------------| 2505/7340 [84:55<163:55, 29.5 steps/min]2025-08-11 16:51:14,650 - agent.ComputerAgent - INFO - LLM processing started with 14 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 14 messages\n", - "\u001b[92m16:51:14 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 16:51:15,336 - agent.ComputerAgent - INFO - LLM processing started with 36 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 36 messages\n", - "\u001b[92m16:51:15 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 34%|█████████████---------------------------| 2505/7340 [84:57<163:58, 29.5 steps/min]2025-08-11 16:51:15,998 - agent.ComputerAgent - INFO - LLM processing started with 16 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 16 messages\n", - "\u001b[92m16:51:16 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 16:51:16,677 - agent.ComputerAgent - INFO - LLM processing started with 10 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 10 messages\n", - "\u001b[92m16:51:16 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 34%|█████████████---------------------------| 2505/7340 [84:58<164:00, 29.5 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 34%|█████████████---------------------------| 2505/7340 [84:59<164:02, 29.5 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v1/environments/c3518cd0-0df6-44e9-8393-0c62002bc984/reset \"HTTP/1.1 200 OK\"\n", - " 34%|█████████████---------------------------| 2505/7340 [85:00<164:04, 29.5 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:51:19 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 16:51:20,733 - agent.ComputerAgent - INFO - Computer: type({'text': 'Stockholm'})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': 'Stockholm'})\n", - " 34%|█████████████---------------------------| 2505/7340 [85:02<164:08, 29.5 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m16:51:21 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Loading checkpoint shards: 0%| | 0/4 [00:00 Slide Pane to toggle it back on.\n", - "- You can resize it by dragging its right edge. If you lose the right-side Properties panel, use View > Sidebar.\n", - "\n", - "Task completed\n", - "INFO:agent.ComputerAgent:Agent: I restored the Slides panel on the left in your LibreOffice Impress.\n", - "\n", - "If it disappears again:\n", - "- Go to the top menu: View > Slide Pane to toggle it back on.\n", - "- You can resize it by dragging its right edge. If you lose the right-side Properties panel, use View > Sidebar.\n", - "\n", - "Task completed\n", - "2025-08-11 17:05:27,005 - agent.ComputerAgent - INFO - Total usage:\n", - " - completion_tokens: 712\n", - " - prompt_tokens: 7355\n", - " - total_tokens: 8067\n", - " - completion_tokens_details:\n", - " - accepted_prediction_tokens: 0\n", - " - audio_tokens: 0\n", - " - reasoning_tokens: 640\n", - " - rejected_prediction_tokens: 0\n", - " - prompt_tokens_details:\n", - " - audio_tokens: 0\n", - " - cached_tokens: 0\n", - " - response_cost: $0.0163\n", - "INFO:agent.ComputerAgent:Total usage:\n", - " - completion_tokens: 712\n", - " - prompt_tokens: 7355\n", - " - total_tokens: 8067\n", - " - completion_tokens_details:\n", - " - accepted_prediction_tokens: 0\n", - " - audio_tokens: 0\n", - " - reasoning_tokens: 640\n", - " - rejected_prediction_tokens: 0\n", - " - prompt_tokens_details:\n", - " - audio_tokens: 0\n", - " - cached_tokens: 0\n", - " - response_cost: $0.0163\n", - " 38%|███████████████-------------------------| 2809/7340 [99:08<159:55, 28.3 steps/min]\u001b[92m17:05:27 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/6f141a0f-f4b0-4f99-b4c4-5217b268c96b/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:05:27 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v1/environments/025be48d-d757-4973-8c17-e42b8f6814b0/reset \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/d2c8050e-87aa-487a-9555-884be2298ade/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:05:28 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:05:28 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m17:05:29 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 17:05:30,349 - agent.ComputerAgent - INFO - Computer: click({'x': 87, 'y': 274})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 87, 'y': 274})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/9b6fc8c3-534a-4e7d-9a9b-4c6bad0e0619/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Loading checkpoint shards: 0%| | 0/4 [00:00 Appearance > Mode = Light\n", - "\n", - "This forces Chrome to use a light theme regardless of your device theme.\n", - "\n", - "If any websites still appear dark, it could be due to:\n", - "- A dark-mode extension (e.g., Dark Reader). Disable it under Extensions.\n", - "- The experimental flag “Auto Dark Mode for Web Contents” being enabled. Visit chrome://flags and ensure it’s Disabled.\n", - "\n", - "Task completed.\n", - "INFO:agent.ComputerAgent:Agent: I’ve turned off Dark Mode in Google Chrome by setting the browser’s appearance to Light:\n", - "- Settings > Appearance > Mode = Light\n", - "\n", - "This forces Chrome to use a light theme regardless of your device theme.\n", - "\n", - "If any websites still appear dark, it could be due to:\n", - "- A dark-mode extension (e.g., Dark Reader). Disable it under Extensions.\n", - "- The experimental flag “Auto Dark Mode for Web Contents” being enabled. Visit chrome://flags and ensure it’s Disabled.\n", - "\n", - "Task completed.\n", - "2025-08-11 17:10:01,595 - agent.ComputerAgent - INFO - Total usage:\n", - " - completion_tokens: 876\n", - " - prompt_tokens: 8770\n", - " - total_tokens: 9646\n", - " - completion_tokens_details:\n", - " - accepted_prediction_tokens: 0\n", - " - audio_tokens: 0\n", - " - reasoning_tokens: 768\n", - " - rejected_prediction_tokens: 0\n", - " - prompt_tokens_details:\n", - " - audio_tokens: 0\n", - " - cached_tokens: 0\n", - " - response_cost: $0.0197\n", - "INFO:agent.ComputerAgent:Total usage:\n", - " - completion_tokens: 876\n", - " - prompt_tokens: 8770\n", - " - total_tokens: 9646\n", - " - completion_tokens_details:\n", - " - accepted_prediction_tokens: 0\n", - " - audio_tokens: 0\n", - " - reasoning_tokens: 768\n", - " - rejected_prediction_tokens: 0\n", - " - prompt_tokens_details:\n", - " - audio_tokens: 0\n", - " - cached_tokens: 0\n", - " - response_cost: $0.0197\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:10:02 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 17:10:02,937 - agent.ComputerAgent - INFO - Computer: click({'x': 400, 'y': 77})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 400, 'y': 77})\n", - "2025-08-11 17:10:03,606 - agent.ComputerAgent - INFO - Computer: double_click({'x': 476, 'y': 392})\n", - "INFO:agent.ComputerAgent:Computer: double_click({'x': 476, 'y': 392})\n", - "\u001b[92m17:10:03 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 17:10:04,225 - agent.ComputerAgent - INFO - LLM processing started with 6 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 6 messages\n", - " 40%|████████████████------------------------| 2945/7340 [103:45<154:51, 28.4 steps/min]\u001b[92m17:10:04 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 17:10:04,897 - agent.ComputerAgent - INFO - Computer: click({'x': 986, 'y': 578})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 986, 'y': 578})\n", - "\u001b[92m17:10:04 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m17:10:05 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - " 40%|████████████████------------------------| 2948/7340 [103:46<154:37, 28.4 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:10:06 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m17:10:06 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 17:10:06,706 - agent.ComputerAgent - INFO - Computer: drag({'path': [{'x': 143, 'y': 188}, {'x': 147, 'y': 229}]})\n", - "INFO:agent.ComputerAgent:Computer: drag({'path': [{'x': 143, 'y': 188}, {'x': 147, 'y': 229}]})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:10:06 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m17:10:07 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/6f141a0f-f4b0-4f99-b4c4-5217b268c96b/invoke \"HTTP/1.1 200 OK\"\n", - " 40%|████████████████------------------------| 2949/7340 [103:49<154:35, 28.4 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 17:10:08,028 - agent.ComputerAgent - INFO - Computer: double_click({'x': 730, 'y': 648})\n", - "INFO:agent.ComputerAgent:Computer: double_click({'x': 730, 'y': 648})\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - " 40%|████████████████------------------------| 2950/7340 [103:50<154:31, 28.4 steps/min]\u001b[92m17:10:08 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 17:10:09,210 - agent.ComputerAgent - INFO - Computer: click({'x': 982, 'y': 167})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 982, 'y': 167})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/6f141a0f-f4b0-4f99-b4c4-5217b268c96b/invoke \"HTTP/1.1 200 OK\"\n", - " 40%|████████████████------------------------| 2960/7340 [103:51<153:40, 28.5 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/091ec079-295e-4528-bad5-f34604d013c2/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/085317e9-3b47-437e-8528-0a0fc0e6e688/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 17:10:10,343 - agent.ComputerAgent - INFO - LLM processing started with 10 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 10 messages\n", - "\u001b[92m17:10:10 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/9b6fc8c3-534a-4e7d-9a9b-4c6bad0e0619/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/9f1aa688-383a-44cc-9959-87e1e32dd55c/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/6f141a0f-f4b0-4f99-b4c4-5217b268c96b/close \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/55d96e84-fdd8-4854-8eca-195f1a84dbc3/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/aa0172ad-f4a9-4f1a-9e06-2d510775dbd0/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/1f48e361-2592-41ee-8818-d6e9174fe800/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 17:10:11,005 - agent.ComputerAgent - INFO - LLM processing started with 8 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 8 messages\n", - "\u001b[92m17:10:11 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/797f1798-0199-4d66-a503-1c5a8d488911/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:10:12 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/88706cb5-896e-4bf5-8b52-5df252945e00/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 40%|████████████████------------------------| 2961/7340 [103:54<153:39, 28.5 steps/min]2025-08-11 17:10:12,983 - agent.ComputerAgent - INFO - LLM processing started with 6 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 6 messages\n", - "\u001b[92m17:10:12 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:10:13 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/ecc77db9-1735-4a1a-ab51-39c4517489fb/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/6b818afb-eeae-43f9-b1ff-7f82844997e2/invoke \"HTTP/1.1 200 OK\"\n", - "Loading checkpoint shards: 0%| | 0/4 [00:00 Options > LibreOffice Impress > General: unchecked “Enable Presenter Console,” then applied and saved.\n", - "\n", - "How to choose which single monitor shows the slideshow (do this once per machine/setup):\n", - "- Slide Show > Slide Show Settings…\n", - "- Under Multiple displays, set “Presentation display” to the monitor you want (Display 1 or Display 2) and keep Presentation mode as “Full screen.”\n", - "- Click OK.\n", - "\n", - "Result:\n", - "- When you start the slideshow, it will fill only the selected monitor.\n", - "- The second monitor will remain free for other work, and the “Current/Next slide” presenter view will no longer appear.\n", - "\n", - "Task completed.\n", - "INFO:agent.ComputerAgent:Agent: I configured LibreOffice Impress to use only one monitor for the slideshow and to disable the Presenter Console.\n", - "\n", - "What I changed:\n", - "- Tools > Options > LibreOffice Impress > General: unchecked “Enable Presenter Console,” then applied and saved.\n", - "\n", - "How to choose which single monitor shows the slideshow (do this once per machine/setup):\n", - "- Slide Show > Slide Show Settings…\n", - "- Under Multiple displays, set “Presentation display” to the monitor you want (Display 1 or Display 2) and keep Presentation mode as “Full screen.”\n", - "- Click OK.\n", - "\n", - "Result:\n", - "- When you start the slideshow, it will fill only the selected monitor.\n", - "- The second monitor will remain free for other work, and the “Current/Next slide” presenter view will no longer appear.\n", - "\n", - "Task completed.\n", - "2025-08-11 17:13:46,540 - agent.ComputerAgent - INFO - Total usage:\n", - " - completion_tokens: 1315\n", - " - prompt_tokens: 13574\n", - " - total_tokens: 14889\n", - " - completion_tokens_details:\n", - " - accepted_prediction_tokens: 0\n", - " - audio_tokens: 0\n", - " - reasoning_tokens: 1152\n", - " - rejected_prediction_tokens: 0\n", - " - prompt_tokens_details:\n", - " - audio_tokens: 0\n", - " - cached_tokens: 9472\n", - " - response_cost: $0.0195\n", - "INFO:agent.ComputerAgent:Total usage:\n", - " - completion_tokens: 1315\n", - " - prompt_tokens: 13574\n", - " - total_tokens: 14889\n", - " - completion_tokens_details:\n", - " - accepted_prediction_tokens: 0\n", - " - audio_tokens: 0\n", - " - reasoning_tokens: 1152\n", - " - rejected_prediction_tokens: 0\n", - " - prompt_tokens_details:\n", - " - audio_tokens: 0\n", - " - cached_tokens: 9472\n", - " - response_cost: $0.0195\n", - " 42%|████████████████------------------------| 3054/7340 [107:28<150:49, 28.4 steps/min]2025-08-11 17:13:47,229 - agent.ComputerAgent - INFO - Computer: click({'x': 969, 'y': 218})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 969, 'y': 218})\n", - "2025-08-11 17:13:47,886 - agent.ComputerAgent - INFO - Computer: double_click({'x': 205, 'y': 214})\n", - "INFO:agent.ComputerAgent:Computer: double_click({'x': 205, 'y': 214})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:13:48 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m17:13:49 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 17:13:50,419 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'ctrl+alt+t'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'ctrl+alt+t'})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - " 42%|████████████████------------------------| 3055/7340 [107:32<150:50, 28.4 steps/min]\u001b[92m17:13:51 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m17:13:51 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m17:13:51 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 17:13:51,700 - agent.ComputerAgent - INFO - LLM processing started with 8 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 8 messages\n", - "\u001b[92m17:13:51 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 17:13:52,387 - agent.ComputerAgent - INFO - Computer: click({'x': 954, 'y': 232})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 954, 'y': 232})\n", - "2025-08-11 17:13:53,035 - agent.ComputerAgent - INFO - Computer: click({'x': 589, 'y': 143})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 589, 'y': 143})\n", - "\u001b[92m17:13:53 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 42%|████████████████------------------------| 3057/7340 [107:34<150:43, 28.4 steps/min]2025-08-11 17:13:53,673 - agent.ComputerAgent - INFO - Computer: scroll({'scroll_y': 660, 'scroll_x': 0, 'x': 658, 'y': 467})\n", - "INFO:agent.ComputerAgent:Computer: scroll({'scroll_y': 660, 'scroll_x': 0, 'x': 658, 'y': 467})\n", - " 42%|████████████████------------------------| 3059/7340 [107:35<150:34, 28.4 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/1f48e361-2592-41ee-8818-d6e9174fe800/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/9b6fc8c3-534a-4e7d-9a9b-4c6bad0e0619/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:13:55 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 42%|████████████████------------------------| 3060/7340 [107:37<150:31, 28.4 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m17:13:55 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 17:13:56,551 - agent.ComputerAgent - INFO - Computer: click({'x': 660, 'y': 104})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 660, 'y': 104})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/1f48e361-2592-41ee-8818-d6e9174fe800/invoke \"HTTP/1.1 200 OK\"\n", - " 42%|████████████████------------------------| 3060/7340 [107:38<150:33, 28.4 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/1f48e361-2592-41ee-8818-d6e9174fe800/close \"HTTP/1.1 200 OK\"\n", - " 42%|████████████████------------------------| 3063/7340 [107:39<150:19, 28.5 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/025be48d-d757-4973-8c17-e42b8f6814b0/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/afb4e623-39bf-4f23-ac18-6c4a71f53c62/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 17:13:58,853 - agent.ComputerAgent - INFO - LLM processing started with 32 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 32 messages\n", - "\u001b[92m17:13:58 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/085317e9-3b47-437e-8528-0a0fc0e6e688/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/9f1aa688-383a-44cc-9959-87e1e32dd55c/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/6bacb467-6eb5-4ead-ac71-a185d2fa5e80/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/55d96e84-fdd8-4854-8eca-195f1a84dbc3/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: GET https://orchestration.hud.so/hud-gym/api/v1/gyms/OSWorld-Ubuntu \"HTTP/1.1 200 OK\"\n", - " 42%|████████████████------------------------| 3063/7340 [107:40<150:21, 28.4 steps/min]2025-08-11 17:13:59,510 - agent.ComputerAgent - INFO - LLM processing started with 20 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 20 messages\n", - "\u001b[92m17:13:59 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 17:14:00,189 - agent.ComputerAgent - INFO - LLM processing started with 28 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 28 messages\n", - "\u001b[92m17:14:00 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 17:14:00,831 - agent.ComputerAgent - INFO - LLM processing started with 36 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 36 messages\n", - "\u001b[92m17:14:00 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/aa0172ad-f4a9-4f1a-9e06-2d510775dbd0/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 42%|████████████████------------------------| 3063/7340 [107:42<150:23, 28.4 steps/min]2025-08-11 17:14:01,477 - agent.ComputerAgent - INFO - LLM processing started with 20 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 20 messages\n", - "\u001b[92m17:14:01 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 17:14:02,169 - agent.ComputerAgent - INFO - LLM processing started with 22 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 22 messages\n", - "\u001b[92m17:14:02 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/f0190121-650c-4779-b26d-2480f313dc84/invoke \"HTTP/1.1 200 OK\"\n", - " 42%|████████████████------------------------| 3063/7340 [107:43<150:25, 28.4 steps/min]2025-08-11 17:14:02,841 - agent.ComputerAgent - INFO - LLM processing started with 24 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 24 messages\n", - "\u001b[92m17:14:02 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 17:14:03,496 - agent.ComputerAgent - INFO - LLM processing started with 26 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 26 messages\n", - "\u001b[92m17:14:03 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 42%|████████████████------------------------| 3063/7340 [107:45<150:27, 28.4 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:14:04 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:14:04 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 17:14:06,224 - agent.ComputerAgent - INFO - Computer: type({'text': 'conda create -n hf python=3.11 --override-channels -c conda-forge -y\\n'})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': 'conda create -n hf python=3.11 --override-channels -c conda-forge -y\\n'})\n", - " 42%|████████████████------------------------| 3064/7340 [107:53<150:33, 28.4 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/091ec079-295e-4528-bad5-f34604d013c2/invoke \"HTTP/1.1 200 OK\"\n", - "Loading checkpoint shards: 50%|█████ | 2/4 [00:03<00:03, 1.58s/it]2025-08-11 17:14:12,505 - agent.ComputerAgent - INFO - LLM processing started with 28 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 28 messages\n", - "\u001b[92m17:14:12 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 42%|████████████████------------------------| 3064/7340 [107:55<150:36, 28.4 steps/min]\u001b[92m17:14:13 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Loading checkpoint shards: 100%|██████████| 4/4 [00:05<00:00, 1.32s/it]\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:14:13 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m17:14:14 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 17:14:15,999 - agent.ComputerAgent - INFO - Computer: type({'text': 'spider-man toys kids'})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': 'spider-man toys kids'})\n", - " 42%|████████████████------------------------| 3064/7340 [107:57<150:40, 28.4 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:14:16 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m17:14:16 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 17:14:17,286 - agent.ComputerAgent - INFO - Computer: click({'button': 'right', 'x': 512, 'y': 384})\n", - "INFO:agent.ComputerAgent:Computer: click({'button': 'right', 'x': 512, 'y': 384})\n", - "\u001b[92m17:14:17 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m17:14:17 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m17:14:17 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 17:14:17,916 - agent.ComputerAgent - INFO - Computer: click({'x': 175, 'y': 183})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 175, 'y': 183})\n", - " 42%|████████████████------------------------| 3065/7340 [107:59<150:37, 28.4 steps/min]2025-08-11 17:14:18,581 - agent.ComputerAgent - INFO - Computer: click({'x': 730, 'y': 275})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 730, 'y': 275})\n", - "2025-08-11 17:14:19,257 - agent.ComputerAgent - INFO - Computer: click({'x': 125, 'y': 182})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 125, 'y': 182})\n", - "\u001b[92m17:14:19 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m17:14:19 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 17:14:19,937 - agent.ComputerAgent - INFO - Computer: click({'x': 184, 'y': 178})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 184, 'y': 178})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:14:20 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 42%|████████████████------------------------| 3067/7340 [108:02<150:31, 28.4 steps/min]2025-08-11 17:14:21,216 - agent.ComputerAgent - INFO - Computer: double_click({'x': 757, 'y': 644})\n", - "INFO:agent.ComputerAgent:Computer: double_click({'x': 757, 'y': 644})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m17:14:21 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:14:21 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:14:22 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 42%|████████████████------------------------| 3070/7340 [108:04<150:19, 28.4 steps/min]\u001b[92m17:14:23 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 17:14:23,849 - agent.ComputerAgent - INFO - Computer: click({'x': 318, 'y': 59})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 318, 'y': 59})\n", - "\u001b[92m17:14:23 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 17:14:25,159 - agent.ComputerAgent - INFO - Computer: wait({})\n", - "INFO:agent.ComputerAgent:Computer: wait({})\n", - "2025-08-11 17:14:25,792 - agent.ComputerAgent - INFO - Computer: click({'x': 910, 'y': 254})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 910, 'y': 254})\n", - "\u001b[92m17:14:25 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/create_environment \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:14:25 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v1/environments/094ee49d-29b5-4911-bfc8-7d0e73a55c44/reset \"HTTP/1.1 200 OK\"\n", - " 42%|████████████████------------------------| 3071/7340 [108:07<150:18, 28.4 steps/min]2025-08-11 17:14:26,480 - agent.ComputerAgent - INFO - Computer: click({'x': 652, 'y': 178})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 652, 'y': 178})\n", - "2025-08-11 17:14:27,163 - agent.ComputerAgent - INFO - Computer: scroll({'scroll_y': 654, 'scroll_x': 0, 'x': 654, 'y': 467})\n", - "INFO:agent.ComputerAgent:Computer: scroll({'scroll_y': 654, 'scroll_x': 0, 'x': 654, 'y': 467})\n", - " 42%|████████████████------------------------| 3076/7340 [108:09<149:56, 28.4 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/6a6179f5-13f9-4283-a0d1-aaafd881b00a/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/9b6fc8c3-534a-4e7d-9a9b-4c6bad0e0619/invoke \"HTTP/1.1 200 OK\"\n", - " 42%|████████████████------------------------| 3078/7340 [108:11<149:49, 28.4 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/create_environment \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v1/environments/6a6179f5-13f9-4283-a0d1-aaafd881b00a/reset \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/9b6fc8c3-534a-4e7d-9a9b-4c6bad0e0619/close \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/094ee49d-29b5-4911-bfc8-7d0e73a55c44/invoke \"HTTP/1.1 200 OK\"\n", - " 42%|████████████████------------------------| 3078/7340 [108:13<149:50, 28.4 steps/min]2025-08-11 17:14:32,090 - agent.ComputerAgent - INFO - LLM processing started with 1 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 1 messages\n", - "\u001b[92m17:14:32 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/55d96e84-fdd8-4854-8eca-195f1a84dbc3/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/025be48d-d757-4973-8c17-e42b8f6814b0/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/085317e9-3b47-437e-8528-0a0fc0e6e688/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/4b18a76d-ef46-4622-9643-9ee6fe4900a3/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/afb4e623-39bf-4f23-ac18-6c4a71f53c62/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/6bacb467-6eb5-4ead-ac71-a185d2fa5e80/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/055e9f8b-8c01-4732-8b5f-ef4fc732f122/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/35bb6fb7-5b34-473c-a541-13215a694bc6/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: GET https://orchestration.hud.so/hud-gym/api/v1/gyms/OSWorld-Ubuntu \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/f0190121-650c-4779-b26d-2480f313dc84/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 17:14:32,771 - agent.ComputerAgent - INFO - LLM processing started with 22 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 22 messages\n", - "\u001b[92m17:14:32 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/9f1aa688-383a-44cc-9959-87e1e32dd55c/invoke \"HTTP/1.1 200 OK\"\n", - " 42%|████████████████------------------------| 3078/7340 [108:14<149:52, 28.4 steps/min]2025-08-11 17:14:33,404 - agent.ComputerAgent - INFO - LLM processing started with 30 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 30 messages\n", - "\u001b[92m17:14:33 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 17:14:34,070 - agent.ComputerAgent - INFO - LLM processing started with 14 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 14 messages\n", - "\u001b[92m17:14:34 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 17:14:34,696 - agent.ComputerAgent - INFO - LLM processing started with 28 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 28 messages\n", - "\u001b[92m17:14:34 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 17:14:35,372 - agent.ComputerAgent - INFO - LLM processing started with 38 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 38 messages\n", - "\u001b[92m17:14:35 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 17:14:36,032 - agent.ComputerAgent - INFO - LLM processing started with 34 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 34 messages\n", - "\u001b[92m17:14:36 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/6a6179f5-13f9-4283-a0d1-aaafd881b00a/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/091ec079-295e-4528-bad5-f34604d013c2/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/96765d66-53fb-41dd-99b6-cd96984e52b3/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/aa0172ad-f4a9-4f1a-9e06-2d510775dbd0/invoke \"HTTP/1.1 200 OK\"\n", - " 42%|████████████████------------------------| 3078/7340 [108:17<149:57, 28.4 steps/min]2025-08-11 17:14:37,022 - agent.ComputerAgent - INFO - LLM processing started with 1 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 1 messages\n", - "\u001b[92m17:14:37 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 17:14:37,680 - agent.ComputerAgent - INFO - LLM processing started with 26 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 26 messages\n", - "\u001b[92m17:14:37 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 42%|████████████████------------------------| 3078/7340 [108:19<149:59, 28.4 steps/min]2025-08-11 17:14:38,337 - agent.ComputerAgent - INFO - LLM processing started with 30 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 30 messages\n", - "\u001b[92m17:14:38 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 17:14:39,001 - agent.ComputerAgent - INFO - LLM processing started with 24 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 24 messages\n", - "\u001b[92m17:14:39 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 17:14:39,701 - agent.ComputerAgent - INFO - LLM processing started with 22 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 22 messages\n", - "\u001b[92m17:14:39 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 42%|████████████████------------------------| 3078/7340 [108:21<150:02, 28.4 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:14:40 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Loading checkpoint shards: 25%|██▌ | 1/4 [00:01<00:04, 1.63s/it] 28.4 steps/min]2025-08-11 17:14:42,801 - agent.ComputerAgent - INFO - LLM processing started with 10 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 10 messages\n", - "\u001b[92m17:14:42 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 42%|████████████████------------------------| 3078/7340 [108:24<150:06, 28.4 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 17:14:43,842 - agent.ComputerAgent - INFO - LLM processing started with 28 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 28 messages\n", - "\u001b[92m17:14:43 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 42%|████████████████------------------------| 3078/7340 [108:25<150:08, 28.4 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "Loading checkpoint shards: 100%|██████████| 4/4 [00:05<00:00, 1.33s/it] 28.4 steps/min]\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:14:47 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 42%|████████████████------------------------| 3078/7340 [108:28<150:12, 28.4 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - " 42%|████████████████------------------------| 3078/7340 [108:29<150:14, 28.4 steps/min]\u001b[92m17:14:48 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 17:14:48,883 - agent.ComputerAgent - INFO - Agent: Taking a screenshot to see the current computer screen.\n", - "INFO:agent.ComputerAgent:Agent: Taking a screenshot to see the current computer screen.\n", - "2025-08-11 17:14:48,885 - agent.ComputerAgent - INFO - Computer: click({'x': 314, 'y': 121})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 314, 'y': 121})\n", - "\u001b[92m17:14:49 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 17:14:49,521 - agent.ComputerAgent - INFO - Computer: double_click({'x': 193, 'y': 178})\n", - "INFO:agent.ComputerAgent:Computer: double_click({'x': 193, 'y': 178})\n", - " 42%|████████████████------------------------| 3080/7340 [108:32<150:07, 28.4 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/create_environment \"HTTP/1.1 200 OK\"\n", - " 42%|████████████████------------------------| 3080/7340 [108:33<150:08, 28.4 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/d351b561-0537-4e9c-84fc-8e1905f2f2c8/invoke \"HTTP/1.1 200 OK\"\n", - " 42%|████████████████------------------------| 3080/7340 [108:34<150:09, 28.4 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 17:14:53,931 - agent.ComputerAgent - INFO - Computer: type({'text': 'Paper Recommendation'})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': 'Paper Recommendation'})\n", - " 42%|████████████████------------------------| 3080/7340 [108:35<150:11, 28.4 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/085317e9-3b47-437e-8528-0a0fc0e6e688/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/094ee49d-29b5-4911-bfc8-7d0e73a55c44/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 17:14:55,591 - agent.ComputerAgent - INFO - LLM processing started with 24 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 24 messages\n", - "\u001b[92m17:14:55 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 42%|████████████████------------------------| 3081/7340 [108:37<150:09, 28.4 steps/min]2025-08-11 17:14:56,259 - agent.ComputerAgent - INFO - LLM processing started with 6 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 6 messages\n", - "\u001b[92m17:14:56 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 17:14:57,587 - agent.ComputerAgent - INFO - Computer: wait({})\n", - "INFO:agent.ComputerAgent:Computer: wait({})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - " 42%|████████████████------------------------| 3081/7340 [108:39<150:12, 28.4 steps/min]\u001b[92m17:14:58 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m17:14:58 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 17:14:59,416 - agent.ComputerAgent - INFO - Computer: click({'x': 107, 'y': 33})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 107, 'y': 33})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 17:15:00,701 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'enter'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'enter'})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - " 42%|████████████████------------------------| 3082/7340 [108:43<150:12, 28.3 steps/min]\u001b[92m17:15:01 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 17:15:02,638 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'ctrl+f'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'ctrl+f'})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/9f1aa688-383a-44cc-9959-87e1e32dd55c/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - " 42%|████████████████------------------------| 3084/7340 [108:45<150:04, 28.4 steps/min]\u001b[92m17:15:03 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m17:15:03 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 17:15:04,289 - agent.ComputerAgent - INFO - LLM processing started with 30 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 30 messages\n", - "\u001b[92m17:15:04 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 17:15:04,939 - agent.ComputerAgent - INFO - Computer: click({'x': 572, 'y': 551})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 572, 'y': 551})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/091ec079-295e-4528-bad5-f34604d013c2/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - " 42%|████████████████------------------------| 3084/7340 [108:47<150:07, 28.3 steps/min]\u001b[92m17:15:05 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m17:15:05 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m17:15:05 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 17:15:06,301 - agent.ComputerAgent - INFO - LLM processing started with 24 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 24 messages\n", - "\u001b[92m17:15:06 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 17:15:07,327 - agent.ComputerAgent - INFO - LLM processing started with 32 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 32 messages\n", - "\u001b[92m17:15:07 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m17:15:07 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 17:15:08,631 - agent.ComputerAgent - INFO - Agent: Taking a screenshot to see the current computer screen.\n", - "INFO:agent.ComputerAgent:Agent: Taking a screenshot to see the current computer screen.\n", - "2025-08-11 17:15:08,632 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'win'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'win'})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:15:09 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 42%|████████████████------------------------| 3085/7340 [108:51<150:07, 28.3 steps/min]2025-08-11 17:15:09,958 - agent.ComputerAgent - INFO - Computer: click({'x': 349, 'y': 207})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 349, 'y': 207})\n", - "\u001b[92m17:15:09 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m17:15:10 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 17:15:11,270 - agent.ComputerAgent - INFO - Computer: drag({'path': [{'x': 147, 'y': 581}, {'x': 147, 'y': 678}]})\n", - "INFO:agent.ComputerAgent:Computer: drag({'path': [{'x': 147, 'y': 581}, {'x': 147, 'y': 678}]})\n", - " 42%|████████████████------------------------| 3086/7340 [108:52<150:05, 28.3 steps/min]\u001b[92m17:15:11 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 17:15:11,906 - agent.ComputerAgent - INFO - Computer: click({'x': 880, 'y': 203})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 880, 'y': 203})\n", - "\u001b[92m17:15:11 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 17:15:12,582 - agent.ComputerAgent - INFO - Computer: click({'x': 918, 'y': 243})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 918, 'y': 243})\n", - " 42%|████████████████------------------------| 3090/7340 [108:55<149:48, 28.4 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:15:14 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 42%|████████████████------------------------| 3090/7340 [108:56<149:50, 28.4 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m17:15:15 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 17:15:15,903 - agent.ComputerAgent - INFO - Computer: click({'x': 894, 'y': 232})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 894, 'y': 232})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/afb4e623-39bf-4f23-ac18-6c4a71f53c62/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/55d96e84-fdd8-4854-8eca-195f1a84dbc3/invoke \"HTTP/1.1 200 OK\"\n", - " 42%|████████████████------------------------| 3090/7340 [108:57<149:51, 28.4 steps/min]2025-08-11 17:15:16,561 - agent.ComputerAgent - INFO - LLM processing started with 26 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 26 messages\n", - "\u001b[92m17:15:16 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 17:15:17,873 - agent.ComputerAgent - INFO - Computer: type({'text': 'conda activate hf\\nconda install -y -c conda-forge datasets\\npython -c \"import datasets, sys; print(\\'datasets version:\\', datasets.__version__)\"\\n'})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': 'conda activate hf\\nconda install -y -c conda-forge datasets\\npython -c \"import datasets, sys; print(\\'datasets version:\\', datasets.__version__)\"\\n'})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/96765d66-53fb-41dd-99b6-cd96984e52b3/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/35bb6fb7-5b34-473c-a541-13215a694bc6/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/094ee49d-29b5-4911-bfc8-7d0e73a55c44/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:15:18 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 17:15:19,817 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'ctrl+z'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'ctrl+z'})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/6bacb467-6eb5-4ead-ac71-a185d2fa5e80/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/f0190121-650c-4779-b26d-2480f313dc84/invoke \"HTTP/1.1 200 OK\"\n", - " 42%|████████████████------------------------| 3091/7340 [109:01<149:52, 28.4 steps/min]2025-08-11 17:15:20,461 - agent.ComputerAgent - INFO - LLM processing started with 32 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 32 messages\n", - "\u001b[92m17:15:20 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 17:15:21,839 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'enter'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'enter'})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/6a6179f5-13f9-4283-a0d1-aaafd881b00a/invoke \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:15:21 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 17:15:22,509 - agent.ComputerAgent - INFO - LLM processing started with 16 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 16 messages\n", - " 42%|████████████████------------------------| 3092/7340 [109:04<149:50, 28.3 steps/min]\u001b[92m17:15:22 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m17:15:22 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 17:15:23,191 - agent.ComputerAgent - INFO - LLM processing started with 30 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 30 messages\n", - "\u001b[92m17:15:23 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 17:15:23,827 - agent.ComputerAgent - INFO - LLM processing started with 40 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 40 messages\n", - "\u001b[92m17:15:23 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m17:15:23 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 42%|████████████████------------------------| 3093/7340 [109:05<149:47, 28.4 steps/min]2025-08-11 17:15:24,498 - agent.ComputerAgent - INFO - Computer: drag({'path': [{'x': 194, 'y': 182}, {'x': 183, 'y': 294}]})\n", - "INFO:agent.ComputerAgent:Computer: drag({'path': [{'x': 194, 'y': 182}, {'x': 183, 'y': 294}]})\n", - "2025-08-11 17:15:25,826 - agent.ComputerAgent - INFO - LLM processing started with 8 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 8 messages\n", - "\u001b[92m17:15:25 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - " 42%|████████████████------------------------| 3093/7340 [109:08<149:51, 28.3 steps/min]\u001b[92m17:15:26 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 17:15:27,861 - agent.ComputerAgent - INFO - Computer: wait({})\n", - "INFO:agent.ComputerAgent:Computer: wait({})\n", - "\u001b[92m17:15:27 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 42%|████████████████------------------------| 3094/7340 [109:09<149:48, 28.3 steps/min]2025-08-11 17:15:28,558 - agent.ComputerAgent - INFO - Computer: click({'x': 205, 'y': 175})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 205, 'y': 175})\n", - "2025-08-11 17:15:29,222 - agent.ComputerAgent - INFO - LLM processing started with 28 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 28 messages\n", - "\u001b[92m17:15:29 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 42%|████████████████------------------------| 3095/7340 [109:11<149:45, 28.3 steps/min]2025-08-11 17:15:29,891 - agent.ComputerAgent - INFO - LLM processing started with 6 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 6 messages\n", - "\u001b[92m17:15:29 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 17:15:30,530 - agent.ComputerAgent - INFO - LLM processing started with 12 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 12 messages\n", - "\u001b[92m17:15:30 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/025be48d-d757-4973-8c17-e42b8f6814b0/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 42%|████████████████------------------------| 3096/7340 [109:12<149:41, 28.4 steps/min]2025-08-11 17:15:31,171 - agent.ComputerAgent - INFO - LLM processing started with 36 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 36 messages\n", - "\u001b[92m17:15:31 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 42%|████████████████------------------------| 3096/7340 [109:13<149:43, 28.3 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/091ec079-295e-4528-bad5-f34604d013c2/invoke \"HTTP/1.1 200 OK\"\n", - " 42%|████████████████------------------------| 3096/7340 [109:14<149:44, 28.3 steps/min]2025-08-11 17:15:32,801 - agent.ComputerAgent - INFO - LLM processing started with 34 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 34 messages\n", - "\u001b[92m17:15:32 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/9f1aa688-383a-44cc-9959-87e1e32dd55c/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 17:15:33,431 - agent.ComputerAgent - INFO - LLM processing started with 26 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 26 messages\n", - "\u001b[92m17:15:33 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 42%|████████████████------------------------| 3096/7340 [109:15<149:45, 28.3 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:15:34 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/085317e9-3b47-437e-8528-0a0fc0e6e688/invoke \"HTTP/1.1 200 OK\"\n", - " 42%|████████████████------------------------| 3096/7340 [109:16<149:47, 28.3 steps/min]\u001b[92m17:15:34 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 17:15:35,248 - agent.ComputerAgent - INFO - Computer: click({'x': 804, 'y': 654})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 804, 'y': 654})\n", - "2025-08-11 17:15:35,931 - agent.ComputerAgent - INFO - LLM processing started with 26 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 26 messages\n", - "\u001b[92m17:15:35 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/aa0172ad-f4a9-4f1a-9e06-2d510775dbd0/invoke \"HTTP/1.1 200 OK\"\n", - " 42%|████████████████------------------------| 3096/7340 [109:17<149:49, 28.3 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 17:15:37,102 - agent.ComputerAgent - INFO - LLM processing started with 32 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 32 messages\n", - "\u001b[92m17:15:37 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/55d96e84-fdd8-4854-8eca-195f1a84dbc3/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 42%|████████████████------------------------| 3097/7340 [109:18<149:45, 28.3 steps/min]2025-08-11 17:15:37,759 - agent.ComputerAgent - INFO - LLM processing started with 28 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 28 messages\n", - "\u001b[92m17:15:37 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 42%|████████████████------------------------| 3097/7340 [109:22<149:51, 28.3 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/094ee49d-29b5-4911-bfc8-7d0e73a55c44/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 17:15:42,613 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'ctrl+p'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'ctrl+p'})\n", - " 42%|████████████████------------------------| 3097/7340 [109:24<149:53, 28.3 steps/min]2025-08-11 17:15:43,243 - agent.ComputerAgent - INFO - LLM processing started with 10 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 10 messages\n", - "\u001b[92m17:15:43 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m17:15:43 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 17:15:44,564 - agent.ComputerAgent - INFO - LLM processing started with 34 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 34 messages\n", - "\u001b[92m17:15:44 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 42%|████████████████------------------------| 3097/7340 [109:26<149:56, 28.3 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m17:15:45 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 17:15:45,740 - agent.ComputerAgent - INFO - Computer: click({'x': 408, 'y': 279})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 408, 'y': 279})\n", - " 42%|████████████████------------------------| 3097/7340 [109:27<149:57, 28.3 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:15:46 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 42%|████████████████------------------------| 3098/7340 [109:28<149:54, 28.3 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m17:15:47 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 17:15:48,097 - agent.ComputerAgent - INFO - Computer: click({'x': 880, 'y': 203})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 880, 'y': 203})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:15:48 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 17:15:50,080 - agent.ComputerAgent - INFO - Computer: type({'text': 'conda install -y -c conda-forge --override-channels datasets\\n'})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': 'conda install -y -c conda-forge --override-channels datasets\\n'})\n", - " 42%|████████████████------------------------| 3098/7340 [109:31<149:58, 28.3 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 17:15:51,375 - agent.ComputerAgent - INFO - Computer: type({'text': 'python --version\\npython3 --version\\nls /usr/bin/python* | head -n 20\\napt-cache policy python4 || apt-cache search python4 | head\\n'})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': 'python --version\\npython3 --version\\nls /usr/bin/python* | head -n 20\\napt-cache policy python4 || apt-cache search python4 | head\\n'})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:15:52 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:15:52 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m17:15:52 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 42%|████████████████------------------------| 3100/7340 [109:34<149:52, 28.3 steps/min]2025-08-11 17:15:53,291 - agent.ComputerAgent - INFO - Computer: move({'x': 914, 'y': 232})\n", - "INFO:agent.ComputerAgent:Computer: move({'x': 914, 'y': 232})\n", - "\u001b[92m17:15:53 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/6a6179f5-13f9-4283-a0d1-aaafd881b00a/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 17:15:53,963 - agent.ComputerAgent - INFO - Computer: click({'x': 935, 'y': 351})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 935, 'y': 351})\n", - "\u001b[92m17:15:53 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m17:15:54 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 42%|████████████████------------------------| 3101/7340 [109:36<149:49, 28.3 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 17:15:55,270 - agent.ComputerAgent - INFO - Computer: click({'x': 225, 'y': 520})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 225, 'y': 520})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m17:15:55 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 42%|████████████████------------------------| 3103/7340 [109:37<149:41, 28.3 steps/min]\u001b[92m17:15:56 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 17:15:56,620 - agent.ComputerAgent - INFO - Computer: click({'x': 235, 'y': 206})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 235, 'y': 206})\n", - "\u001b[92m17:15:56 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 17:15:57,299 - agent.ComputerAgent - INFO - Computer: scroll({'scroll_y': 659, 'scroll_x': 0, 'x': 840, 'y': 467})\n", - "INFO:agent.ComputerAgent:Computer: scroll({'scroll_y': 659, 'scroll_x': 0, 'x': 840, 'y': 467})\n", - " 42%|████████████████------------------------| 3104/7340 [109:39<149:38, 28.3 steps/min]2025-08-11 17:15:57,924 - agent.ComputerAgent - INFO - LLM processing started with 8 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 8 messages\n", - "\u001b[92m17:15:57 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:15:58 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 42%|████████████████------------------------| 3106/7340 [109:40<149:30, 28.3 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m17:15:59 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m17:15:59 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m17:15:59 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 17:16:00,272 - agent.ComputerAgent - INFO - Computer: drag({'path': [{'x': 193, 'y': 180}, {'x': 184, 'y': 293}]})\n", - "INFO:agent.ComputerAgent:Computer: drag({'path': [{'x': 193, 'y': 180}, {'x': 184, 'y': 293}]})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - " 42%|████████████████------------------------| 3106/7340 [109:42<149:33, 28.3 steps/min]\u001b[92m17:16:00 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m17:16:01 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 17:16:02,072 - agent.ComputerAgent - INFO - Computer: click({'x': 1011, 'y': 62})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 1011, 'y': 62})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/6bacb467-6eb5-4ead-ac71-a185d2fa5e80/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/f0190121-650c-4779-b26d-2480f313dc84/invoke \"HTTP/1.1 200 OK\"\n", - " 42%|████████████████------------------------| 3107/7340 [109:43<149:29, 28.3 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/091ec079-295e-4528-bad5-f34604d013c2/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/094ee49d-29b5-4911-bfc8-7d0e73a55c44/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 17:16:02,765 - agent.ComputerAgent - INFO - LLM processing started with 42 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 42 messages\n", - "\u001b[92m17:16:02 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/96765d66-53fb-41dd-99b6-cd96984e52b3/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/025be48d-d757-4973-8c17-e42b8f6814b0/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 17:16:03,441 - agent.ComputerAgent - INFO - LLM processing started with 36 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 36 messages\n", - "\u001b[92m17:16:03 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 17:16:04,480 - agent.ComputerAgent - INFO - LLM processing started with 30 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 30 messages\n", - "\u001b[92m17:16:04 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/55d96e84-fdd8-4854-8eca-195f1a84dbc3/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/create_environment \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/aa0172ad-f4a9-4f1a-9e06-2d510775dbd0/invoke \"HTTP/1.1 200 OK\"\n", - " 42%|████████████████------------------------| 3108/7340 [109:46<149:28, 28.3 steps/min]2025-08-11 17:16:05,143 - agent.ComputerAgent - INFO - LLM processing started with 38 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 38 messages\n", - "\u001b[92m17:16:05 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 17:16:05,823 - agent.ComputerAgent - INFO - LLM processing started with 12 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 12 messages\n", - "\u001b[92m17:16:05 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/085317e9-3b47-437e-8528-0a0fc0e6e688/invoke \"HTTP/1.1 200 OK\"\n", - " 42%|████████████████------------------------| 3108/7340 [109:47<149:29, 28.3 steps/min]2025-08-11 17:16:06,914 - agent.ComputerAgent - INFO - LLM processing started with 30 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 30 messages\n", - "\u001b[92m17:16:06 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/dc026dd3-8d59-43e0-a475-ecef72f1db12/invoke \"HTTP/1.1 200 OK\"\n", - " 42%|████████████████------------------------| 3108/7340 [109:48<149:31, 28.3 steps/min]2025-08-11 17:16:07,563 - agent.ComputerAgent - INFO - LLM processing started with 34 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 34 messages\n", - "\u001b[92m17:16:07 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 17:16:08,252 - agent.ComputerAgent - INFO - LLM processing started with 28 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 28 messages\n", - "\u001b[92m17:16:08 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:16:09 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 42%|████████████████------------------------| 3108/7340 [109:50<149:34, 28.3 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:16:09 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 17:16:10,625 - agent.ComputerAgent - INFO - LLM processing started with 14 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 14 messages\n", - "\u001b[92m17:16:10 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m17:16:10 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/afb4e623-39bf-4f23-ac18-6c4a71f53c62/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 42%|████████████████------------------------| 3108/7340 [109:52<149:36, 28.3 steps/min]2025-08-11 17:16:11,298 - agent.ComputerAgent - INFO - Computer: click({'x': 422, 'y': 249})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 422, 'y': 249})\n", - "\u001b[92m17:16:11 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 17:16:11,982 - agent.ComputerAgent - INFO - LLM processing started with 36 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 36 messages\n", - "\u001b[92m17:16:12 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 17:16:12,642 - agent.ComputerAgent - INFO - Computer: click({'x': 381, 'y': 91})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 381, 'y': 91})\n", - " 42%|████████████████------------------------| 3110/7340 [109:55<149:30, 28.3 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 17:16:15,007 - agent.ComputerAgent - INFO - Computer: wait({})\n", - "INFO:agent.ComputerAgent:Computer: wait({})\n", - " 42%|████████████████------------------------| 3111/7340 [109:57<149:28, 28.3 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:16:16 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:16:17 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 42%|████████████████------------------------| 3111/7340 [109:59<149:30, 28.3 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m17:16:17 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 17:16:18,043 - agent.ComputerAgent - INFO - Computer: click({'x': 413, 'y': 587})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 413, 'y': 587})\n", - "\u001b[92m17:16:18 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 17:16:18,709 - agent.ComputerAgent - INFO - Computer: click({'x': 125, 'y': 182})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 125, 'y': 182})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/9f1aa688-383a-44cc-9959-87e1e32dd55c/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/6a6179f5-13f9-4283-a0d1-aaafd881b00a/invoke \"HTTP/1.1 200 OK\"\n", - " 42%|████████████████------------------------| 3111/7340 [110:00<149:32, 28.3 steps/min]2025-08-11 17:16:19,335 - agent.ComputerAgent - INFO - LLM processing started with 28 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 28 messages\n", - "\u001b[92m17:16:19 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 17:16:20,773 - agent.ComputerAgent - INFO - LLM processing started with 10 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 10 messages\n", - "\u001b[92m17:16:20 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - " 42%|████████████████------------------------| 3113/7340 [110:03<149:26, 28.3 steps/min]\u001b[92m17:16:21 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/091ec079-295e-4528-bad5-f34604d013c2/invoke \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 17:16:22,077 - agent.ComputerAgent - INFO - LLM processing started with 38 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 38 messages\n", - "\u001b[92m17:16:22 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m17:16:22 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 17:16:23,083 - agent.ComputerAgent - INFO - Computer: click({'x': 839, 'y': 234})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 839, 'y': 234})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:16:23 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m17:16:24 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v1/environments/730002fc-5760-41b0-97b8-f6783353a242/reset \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 42%|████████████████------------------------| 3113/7340 [110:06<149:30, 28.3 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m17:16:25 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/055e9f8b-8c01-4732-8b5f-ef4fc732f122/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m17:16:25 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/094ee49d-29b5-4911-bfc8-7d0e73a55c44/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 17:16:25,728 - agent.ComputerAgent - INFO - Computer: click({'x': 925, 'y': 244})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 925, 'y': 244})\n", - "\u001b[92m17:16:25 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 42%|████████████████------------------------| 3114/7340 [110:07<149:26, 28.3 steps/min]2025-08-11 17:16:26,375 - agent.ComputerAgent - INFO - LLM processing started with 18 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 18 messages\n", - "\u001b[92m17:16:26 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 17:16:27,003 - agent.ComputerAgent - INFO - Computer: click({'x': 847, 'y': 404})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 847, 'y': 404})\n", - "\u001b[92m17:16:27 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 42%|████████████████------------------------| 3115/7340 [110:08<149:23, 28.3 steps/min]2025-08-11 17:16:28,017 - agent.ComputerAgent - INFO - Computer: click({'x': 880, 'y': 203})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 880, 'y': 203})\n", - "2025-08-11 17:16:28,682 - agent.ComputerAgent - INFO - LLM processing started with 14 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 14 messages\n", - "\u001b[92m17:16:28 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 42%|████████████████------------------------| 3116/7340 [110:10<149:21, 28.3 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 17:16:30,033 - agent.ComputerAgent - INFO - Computer: wait({})\n", - "INFO:agent.ComputerAgent:Computer: wait({})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:16:30 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 42%|████████████████------------------------| 3117/7340 [110:12<149:18, 28.3 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/730002fc-5760-41b0-97b8-f6783353a242/invoke \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:16:31 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 17:16:31,839 - agent.ComputerAgent - INFO - Computer: click({'x': 131, 'y': 181})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 131, 'y': 181})\n", - " 42%|████████████████------------------------| 3118/7340 [110:13<149:15, 28.3 steps/min]2025-08-11 17:16:32,503 - agent.ComputerAgent - INFO - LLM processing started with 1 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 1 messages\n", - "\u001b[92m17:16:32 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 42%|████████████████------------------------| 3119/7340 [110:14<149:11, 28.3 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/55d96e84-fdd8-4854-8eca-195f1a84dbc3/invoke \"HTTP/1.1 200 OK\"\n", - " 42%|████████████████------------------------| 3119/7340 [110:15<149:12, 28.3 steps/min]2025-08-11 17:16:34,177 - agent.ComputerAgent - INFO - LLM processing started with 32 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 32 messages\n", - "\u001b[92m17:16:34 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/6bacb467-6eb5-4ead-ac71-a185d2fa5e80/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:16:35 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 17:16:36,680 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'ctrl+f'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'ctrl+f'})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/afb4e623-39bf-4f23-ac18-6c4a71f53c62/invoke \"HTTP/1.1 200 OK\"\n", - " 42%|████████████████------------------------| 3119/7340 [110:18<149:16, 28.3 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/f0190121-650c-4779-b26d-2480f313dc84/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m17:16:37 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/091ec079-295e-4528-bad5-f34604d013c2/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m17:16:37 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/35bb6fb7-5b34-473c-a541-13215a694bc6/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 17:16:37,927 - agent.ComputerAgent - INFO - LLM processing started with 36 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 36 messages\n", - "\u001b[92m17:16:37 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 17:16:38,544 - agent.ComputerAgent - INFO - LLM processing started with 38 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 38 messages\n", - "\u001b[92m17:16:38 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 17:16:39,208 - agent.ComputerAgent - INFO - Computer: click({'x': 940, 'y': 202})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 940, 'y': 202})\n", - "\u001b[92m17:16:39 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:16:39 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 42%|████████████████------------------------| 3119/7340 [110:21<149:21, 28.3 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 17:16:40,538 - agent.ComputerAgent - INFO - Computer: scroll({'scroll_y': 600, 'x': 422, 'y': 249})\n", - "INFO:agent.ComputerAgent:Computer: scroll({'scroll_y': 600, 'x': 422, 'y': 249})\n", - "2025-08-11 17:16:41,185 - agent.ComputerAgent - INFO - LLM processing started with 40 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 40 messages\n", - "\u001b[92m17:16:41 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/6bacb467-6eb5-4ead-ac71-a185d2fa5e80/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - " 43%|█████████████████-----------------------| 3121/7340 [110:23<149:14, 28.3 steps/min]\u001b[92m17:16:42 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 17:16:42,861 - agent.ComputerAgent - INFO - LLM processing started with 32 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 32 messages\n", - "\u001b[92m17:16:42 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:16:42 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 17:16:44,179 - agent.ComputerAgent - INFO - Computer: click({'x': 115, 'y': 184})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 115, 'y': 184})\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 17:16:44,861 - agent.ComputerAgent - INFO - Computer: click({'x': 185, 'y': 177})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 185, 'y': 177})\n", - " 43%|█████████████████-----------------------| 3121/7340 [110:26<149:17, 28.3 steps/min]\u001b[92m17:16:44 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 17:16:45,524 - agent.ComputerAgent - INFO - Agent: Taking a screenshot to see the current computer screen.\n", - "INFO:agent.ComputerAgent:Agent: Taking a screenshot to see the current computer screen.\n", - "2025-08-11 17:16:45,525 - agent.ComputerAgent - INFO - Computer: click({'x': 345, 'y': 202})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 345, 'y': 202})\n", - "2025-08-11 17:16:46,155 - agent.ComputerAgent - INFO - LLM processing started with 32 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 32 messages\n", - "\u001b[92m17:16:46 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 43%|█████████████████-----------------------| 3123/7340 [110:27<149:09, 28.3 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:16:46 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 17:16:48,493 - agent.ComputerAgent - INFO - Computer: wait({})\n", - "INFO:agent.ComputerAgent:Computer: wait({})\n", - "\u001b[92m17:16:48 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/6bacb467-6eb5-4ead-ac71-a185d2fa5e80/close \"HTTP/1.1 200 OK\"\n", - " 43%|█████████████████-----------------------| 3124/7340 [110:30<149:07, 28.3 steps/min]2025-08-11 17:16:49,152 - agent.ComputerAgent - INFO - Computer: click({'x': 964, 'y': 734})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 964, 'y': 734})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/6a6179f5-13f9-4283-a0d1-aaafd881b00a/invoke \"HTTP/1.1 200 OK\"\n", - " 43%|█████████████████-----------------------| 3125/7340 [110:31<149:04, 28.3 steps/min]" - ] - }, - { - "name": "stdout", - "output_type": "stream", - "text": [ - "No screenshot found, taking screenshot\n" - ] - }, - { - "name": "stderr", - "output_type": "stream", - "text": [ - "2025-08-11 17:16:50,459 - agent.ComputerAgent - INFO - LLM processing started with 13 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 13 messages\n", - "\u001b[92m17:16:50 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 43%|█████████████████-----------------------| 3126/7340 [110:32<149:01, 28.3 steps/min]INFO:httpx:HTTP Request: GET https://orchestration.hud.so/hud-gym/api/v1/gyms/OSWorld-Ubuntu \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/055e9f8b-8c01-4732-8b5f-ef4fc732f122/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:16:51 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/025be48d-d757-4973-8c17-e42b8f6814b0/invoke \"HTTP/1.1 200 OK\"\n", - "Loading checkpoint shards: 0%| | 0/4 [00:00/dev/null || echo 'no python4 found'\\n\"})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': \"which python4 || command -v python4 || ls /usr/bin/python4* 2>/dev/null || echo 'no python4 found'\\n\"})\n", - " 43%|█████████████████-----------------------| 3141/7340 [111:10<148:37, 28.3 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 400 Bad Request\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/ae2379a3-a039-4954-afc2-582f8ebffdd2/invoke \"HTTP/1.1 200 OK\"\n", - " 43%|█████████████████-----------------------| 3143/7340 [111:11<148:28, 28.3 steps/min]2025-08-11 17:17:30,119 - agent.ComputerAgent - INFO - LLM processing started with 6 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 6 messages\n", - "\u001b[92m17:17:30 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/091ec079-295e-4528-bad5-f34604d013c2/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/6a6179f5-13f9-4283-a0d1-aaafd881b00a/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:17:30 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 17:17:32,137 - agent.ComputerAgent - INFO - Computer: type({'text': '=DATEDIF(DATEVALUE(REGEX(A2;\".* \";\"\"));TODAY();\"y\")'})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': '=DATEDIF(DATEVALUE(REGEX(A2;\".* \";\"\"));TODAY();\"y\")'})\n", - "2025-08-11 17:17:32,777 - agent.ComputerAgent - INFO - LLM processing started with 23 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 23 messages\n", - "\u001b[92m17:17:32 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 17:17:34,149 - agent.ComputerAgent - INFO - Computer: type({'text': 'LLM Powered Autonomous Agents.pdf'})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': 'LLM Powered Autonomous Agents.pdf'})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:17:34 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 43%|█████████████████-----------------------| 3143/7340 [111:17<148:36, 28.2 steps/min]\u001b[92m17:17:35 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:17:35 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m17:17:36 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/96765d66-53fb-41dd-99b6-cd96984e52b3/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 17:17:36,722 - agent.ComputerAgent - INFO - Computer: click({'x': 488, 'y': 368})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 488, 'y': 368})\n", - "\u001b[92m17:17:36 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m17:17:36 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 43%|█████████████████-----------------------| 3145/7340 [111:18<148:28, 28.3 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 17:17:37,353 - agent.ComputerAgent - INFO - Computer: click({'x': 349, 'y': 201})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 349, 'y': 201})\n", - "2025-08-11 17:17:37,994 - agent.ComputerAgent - INFO - Computer: scroll({'scroll_y': 660, 'scroll_x': 0, 'x': 706, 'y': 659})\n", - "INFO:agent.ComputerAgent:Computer: scroll({'scroll_y': 660, 'scroll_x': 0, 'x': 706, 'y': 659})\n", - "\u001b[92m17:17:37 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 400 Bad Request\"\n", - " 43%|█████████████████-----------------------| 3147/7340 [111:19<148:19, 28.3 steps/min]\u001b[92m17:17:38 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 17:17:38,649 - agent.ComputerAgent - INFO - LLM processing started with 18 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 18 messages\n", - "\u001b[92m17:17:38 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m17:17:38 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:17:39 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 17:17:40,658 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'ctrl+a'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'ctrl+a'})\n", - "2025-08-11 17:17:41,329 - agent.ComputerAgent - INFO - Computer: drag({'path': [{'x': 194, 'y': 183}, {'x': 184, 'y': 291}]})\n", - "INFO:agent.ComputerAgent:Computer: drag({'path': [{'x': 194, 'y': 183}, {'x': 184, 'y': 291}]})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/091ec079-295e-4528-bad5-f34604d013c2/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/6a6179f5-13f9-4283-a0d1-aaafd881b00a/invoke \"HTTP/1.1 200 OK\"\n", - " 43%|█████████████████-----------------------| 3149/7340 [111:23<148:14, 28.3 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 17:17:41,970 - agent.ComputerAgent - INFO - LLM processing started with 25 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 25 messages\n", - "\u001b[92m17:17:41 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 17:17:42,600 - agent.ComputerAgent - INFO - LLM processing started with 8 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 8 messages\n", - "\u001b[92m17:17:42 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m17:17:42 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/afb4e623-39bf-4f23-ac18-6c4a71f53c62/invoke \"HTTP/1.1 200 OK\"\n", - " 43%|█████████████████-----------------------| 3150/7340 [111:24<148:11, 28.3 steps/min]2025-08-11 17:17:43,277 - agent.ComputerAgent - INFO - Computer: click({'x': 850, 'y': 202})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 850, 'y': 202})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/055e9f8b-8c01-4732-8b5f-ef4fc732f122/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/730002fc-5760-41b0-97b8-f6783353a242/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/9f1aa688-383a-44cc-9959-87e1e32dd55c/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 17:17:43,930 - agent.ComputerAgent - INFO - LLM processing started with 42 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 42 messages\n", - "\u001b[92m17:17:43 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 43%|█████████████████-----------------------| 3150/7340 [111:25<148:13, 28.3 steps/min]2025-08-11 17:17:44,570 - agent.ComputerAgent - INFO - LLM processing started with 10 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 10 messages\n", - "\u001b[92m17:17:44 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 17:17:45,261 - agent.ComputerAgent - INFO - LLM processing started with 22 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 22 messages\n", - "\u001b[92m17:17:45 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 17:17:45,939 - agent.ComputerAgent - INFO - LLM processing started with 32 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 32 messages\n", - "\u001b[92m17:17:45 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/091ec079-295e-4528-bad5-f34604d013c2/close \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - " 43%|█████████████████-----------------------| 3151/7340 [111:27<148:10, 28.3 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 400 Bad Request\"\n", - " 43%|█████████████████-----------------------| 3152/7340 [111:31<148:11, 28.3 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/aa0172ad-f4a9-4f1a-9e06-2d510775dbd0/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/085317e9-3b47-437e-8528-0a0fc0e6e688/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 17:17:51,370 - agent.ComputerAgent - INFO - LLM processing started with 34 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 34 messages\n", - "\u001b[92m17:17:51 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: GET https://orchestration.hud.so/hud-gym/api/v1/gyms/OSWorld-Ubuntu \"HTTP/1.1 200 OK\"\n", - " 43%|█████████████████-----------------------| 3152/7340 [111:33<148:13, 28.3 steps/min]\u001b[92m17:17:52 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/6a6179f5-13f9-4283-a0d1-aaafd881b00a/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 17:17:52,759 - agent.ComputerAgent - INFO - LLM processing started with 27 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 27 messages\n", - "\u001b[92m17:17:52 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 43%|█████████████████-----------------------| 3152/7340 [111:34<148:14, 28.3 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 17:17:53,422 - agent.ComputerAgent - INFO - LLM processing started with 40 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 40 messages\n", - "\u001b[92m17:17:53 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m17:17:54 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 43%|█████████████████-----------------------| 3152/7340 [111:36<148:18, 28.2 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 400 Bad Request\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:17:55 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 43%|█████████████████-----------------------| 3153/7340 [111:37<148:14, 28.2 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/f0190121-650c-4779-b26d-2480f313dc84/invoke \"HTTP/1.1 200 OK\"\n", - "Loading checkpoint shards: 25%|██▌ | 1/4 [00:01<00:05, 1.77s/it]2025-08-11 17:17:57,266 - agent.ComputerAgent - INFO - LLM processing started with 36 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 36 messages\n", - "\u001b[92m17:17:57 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/6a6179f5-13f9-4283-a0d1-aaafd881b00a/invoke \"HTTP/1.1 200 OK\"\n", - " 43%|█████████████████-----------------------| 3153/7340 [111:39<148:15, 28.2 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 17:17:57,933 - agent.ComputerAgent - INFO - LLM processing started with 29 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 29 messages\n", - "\u001b[92m17:17:57 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "Loading checkpoint shards: 100%|██████████| 4/4 [00:05<00:00, 1.39s/it] 28.2 steps/min]\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/094ee49d-29b5-4911-bfc8-7d0e73a55c44/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:18:01 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 400 Bad Request\"\n", - " 43%|█████████████████-----------------------| 3154/7340 [111:43<148:16, 28.2 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m17:18:02 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m17:18:02 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m17:18:02 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/6a6179f5-13f9-4283-a0d1-aaafd881b00a/invoke \"HTTP/1.1 200 OK\"\n", - " 43%|█████████████████-----------------------| 3167/7340 [111:44<147:14, 28.3 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 17:18:03,401 - agent.ComputerAgent - INFO - LLM processing started with 31 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 31 messages\n", - "\u001b[92m17:18:03 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 17:18:04,060 - agent.ComputerAgent - INFO - Computer: click({'x': 666, 'y': 219})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 666, 'y': 219})\n", - "\u001b[92m17:18:04 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m17:18:04 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m17:18:04 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/094ee49d-29b5-4911-bfc8-7d0e73a55c44/close \"HTTP/1.1 200 OK\"\n", - " 43%|█████████████████-----------------------| 3167/7340 [111:45<147:15, 28.3 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 17:18:04,733 - agent.ComputerAgent - INFO - Computer: click({'x': 442, 'y': 162})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 442, 'y': 162})\n", - "2025-08-11 17:18:05,375 - agent.ComputerAgent - INFO - Computer: click({'x': 811, 'y': 336})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 811, 'y': 336})\n", - "2025-08-11 17:18:06,055 - agent.ComputerAgent - INFO - Computer: double_click({'x': 347, 'y': 222})\n", - "INFO:agent.ComputerAgent:Computer: double_click({'x': 347, 'y': 222})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 400 Bad Request\"\n", - "\u001b[92m17:18:06 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m17:18:06 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 43%|█████████████████-----------------------| 3169/7340 [111:48<147:09, 28.3 steps/min]2025-08-11 17:18:07,401 - agent.ComputerAgent - INFO - Computer: click({'x': 536, 'y': 276})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 536, 'y': 276})\n", - "2025-08-11 17:18:08,031 - agent.ComputerAgent - INFO - Computer: double_click({'x': 489, 'y': 368})\n", - "INFO:agent.ComputerAgent:Computer: double_click({'x': 489, 'y': 368})\n", - " 43%|█████████████████-----------------------| 3172/7340 [111:49<146:56, 28.4 steps/min]INFO:httpx:HTTP Request: GET https://orchestration.hud.so/hud-gym/api/v1/gyms/OSWorld-Ubuntu \"HTTP/1.1 200 OK\"\n", - " 43%|█████████████████-----------------------| 3174/7340 [111:50<146:48, 28.4 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/6a6179f5-13f9-4283-a0d1-aaafd881b00a/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 17:18:09,694 - agent.ComputerAgent - INFO - LLM processing started with 33 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 33 messages\n", - "\u001b[92m17:18:09 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 43%|█████████████████-----------------------| 3174/7340 [111:52<146:50, 28.4 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 400 Bad Request\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/create_environment \"HTTP/1.1 200 OK\"\n", - " 43%|█████████████████-----------------------| 3175/7340 [111:53<146:47, 28.4 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 17:18:13,931 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'ctrl+c'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'ctrl+c'})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:18:14 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/9f1aa688-383a-44cc-9959-87e1e32dd55c/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/6a6179f5-13f9-4283-a0d1-aaafd881b00a/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/730002fc-5760-41b0-97b8-f6783353a242/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/55d96e84-fdd8-4854-8eca-195f1a84dbc3/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/ae2379a3-a039-4954-afc2-582f8ebffdd2/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/afb4e623-39bf-4f23-ac18-6c4a71f53c62/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:18:15 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/1909e6f5-b395-4e1d-b1f7-b06406f8731b/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/025be48d-d757-4973-8c17-e42b8f6814b0/invoke \"HTTP/1.1 200 OK\"\n", - " 43%|█████████████████-----------------------| 3175/7340 [111:56<146:51, 28.4 steps/min]2025-08-11 17:18:15,929 - agent.ComputerAgent - INFO - LLM processing started with 35 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 35 messages\n", - "\u001b[92m17:18:15 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 17:18:16,592 - agent.ComputerAgent - INFO - LLM processing started with 36 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 36 messages\n", - "\u001b[92m17:18:16 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v1/environments/4b18a76d-ef46-4622-9643-9ee6fe4900a3/reset \"HTTP/1.1 200 OK\"\n", - " 43%|█████████████████-----------------------| 3175/7340 [111:58<146:53, 28.4 steps/min]2025-08-11 17:18:17,271 - agent.ComputerAgent - INFO - LLM processing started with 34 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 34 messages\n", - "\u001b[92m17:18:17 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 17:18:17,933 - agent.ComputerAgent - INFO - LLM processing started with 10 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 10 messages\n", - "\u001b[92m17:18:17 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 43%|█████████████████-----------------------| 3175/7340 [111:59<146:55, 28.3 steps/min]2025-08-11 17:18:18,584 - agent.ComputerAgent - INFO - LLM processing started with 12 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 12 messages\n", - "\u001b[92m17:18:18 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v1/environments/b6bbc5bc-5598-4043-be1e-6ebf2da5f046/reset \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Loading checkpoint shards: 0%| | 0/4 [00:00 arr[j + 1]:\\n arr[j], arr[j + 1] = arr[j + 1], arr[j]\\n swapped = True\\n if not swapped:\\n break'})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': '\\n n = len(arr)\\n for i in range(n):\\n swapped = False\\n for j in range(0, n - i - 1):\\n if arr[j] > arr[j + 1]:\\n arr[j], arr[j + 1] = arr[j + 1], arr[j]\\n swapped = True\\n if not swapped:\\n break'})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/create_environment \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/a4b4d291-1fca-4038-8670-448014a55182/invoke \"HTTP/1.1 200 OK\"\n", - " 47%|██████████████████----------------------| 3439/7340 [120:46<137:00, 28.5 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/b3bde0e0-c60f-4177-b7dd-15e361558126/invoke \"HTTP/1.1 200 OK\"\n", - " 47%|██████████████████----------------------| 3439/7340 [120:47<137:01, 28.5 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:27:06 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 17:27:07,882 - agent.ComputerAgent - INFO - Computer: type({'text': 'libreoffice --headless --convert-to pdf --outdir . -- *.doc\\n'})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': 'libreoffice --headless --convert-to pdf --outdir . -- *.doc\\n'})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:27:07 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m17:27:08 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/dc026dd3-8d59-43e0-a475-ecef72f1db12/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/b6bbc5bc-5598-4043-be1e-6ebf2da5f046/invoke \"HTTP/1.1 200 OK\"\n", - " 47%|██████████████████----------------------| 3439/7340 [120:50<137:04, 28.5 steps/min]2025-08-11 17:27:09,249 - agent.ComputerAgent - INFO - Agent: Taking a screenshot to see the current computer screen.\n", - "INFO:agent.ComputerAgent:Agent: Taking a screenshot to see the current computer screen.\n", - "2025-08-11 17:27:09,250 - agent.ComputerAgent - INFO - Computer: double_click({'x': 984, 'y': 491})\n", - "INFO:agent.ComputerAgent:Computer: double_click({'x': 984, 'y': 491})\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 17:27:09,926 - agent.ComputerAgent - INFO - LLM processing started with 28 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 28 messages\n", - "\u001b[92m17:27:09 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m17:27:10 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/94463065-a78e-479a-b964-45ad23a48cbb/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:27:10 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 47%|██████████████████----------------------| 3440/7340 [120:52<137:02, 28.5 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 17:27:11,254 - agent.ComputerAgent - INFO - Computer: click({'x': 153, 'y': 53})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 153, 'y': 53})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 17:27:12,616 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'ctrl+s'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'ctrl+s'})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/create_environment \"HTTP/1.1 200 OK\"\n", - "2025-08-11 17:27:13,257 - agent.ComputerAgent - INFO - LLM processing started with 36 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 36 messages\n", - "\u001b[92m17:27:13 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m17:27:13 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 17:27:14,573 - agent.ComputerAgent - INFO - Computer: type({'text': 'do not track'})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': 'do not track'})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 17:27:15,903 - agent.ComputerAgent - INFO - Computer: type({'text': '30'})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': '30'})\n", - " 47%|██████████████████----------------------| 3441/7340 [120:57<137:03, 28.4 steps/min]2025-08-11 17:27:16,549 - agent.ComputerAgent - INFO - LLM processing started with 18 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 18 messages\n", - "\u001b[92m17:27:16 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 17:27:17,244 - agent.ComputerAgent - INFO - Computer: click({'x': 268, 'y': 329})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 268, 'y': 329})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - " 47%|██████████████████----------------------| 3444/7340 [120:59<136:52, 28.5 steps/min]\u001b[92m17:27:17 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 17:27:18,557 - agent.ComputerAgent - INFO - LLM processing started with 20 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 20 messages\n", - "\u001b[92m17:27:18 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m17:27:18 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 17:27:19,249 - agent.ComputerAgent - INFO - Computer: click({'x': 955, 'y': 130})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 955, 'y': 130})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/127b9298-d3cc-4b90-8567-e45146efa729/invoke \"HTTP/1.1 200 OK\"\n", - " 47%|██████████████████----------------------| 3445/7340 [121:00<136:49, 28.5 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:27:20 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 47%|██████████████████----------------------| 3446/7340 [121:02<136:46, 28.5 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m17:27:20 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 17:27:21,579 - agent.ComputerAgent - INFO - Computer: click({'x': 188, 'y': 105})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 188, 'y': 105})\n", - " 47%|██████████████████----------------------| 3446/7340 [121:03<136:47, 28.5 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/f8984906-7392-4305-88fa-ae9a4808fa8d/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/050a0934-63e8-46a0-8868-de32b28174ef/invoke \"HTTP/1.1 200 OK\"\n", - " 47%|██████████████████----------------------| 3447/7340 [121:04<136:44, 28.5 steps/min]2025-08-11 17:27:23,268 - agent.ComputerAgent - INFO - LLM processing started with 32 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 32 messages\n", - "\u001b[92m17:27:23 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/57944bbf-74a1-4e6d-9401-f7b0144460f7/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/b28701c2-0fa4-4b07-bace-735fd2133893/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 17:27:23,935 - agent.ComputerAgent - INFO - LLM processing started with 10 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 10 messages\n", - "\u001b[92m17:27:23 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/0322563b-daf3-41ae-8a08-f5ecd9282bcc/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/792a6953-2092-47e4-a8a8-57a4af4e3be1/invoke \"HTTP/1.1 200 OK\"\n", - " 47%|██████████████████----------------------| 3447/7340 [121:05<136:45, 28.5 steps/min]2025-08-11 17:27:24,567 - agent.ComputerAgent - INFO - LLM processing started with 24 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 24 messages\n", - "\u001b[92m17:27:24 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 17:27:25,196 - agent.ComputerAgent - INFO - LLM processing started with 24 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 24 messages\n", - "\u001b[92m17:27:25 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 47%|██████████████████----------------------| 3447/7340 [121:07<136:47, 28.5 steps/min]2025-08-11 17:27:25,887 - agent.ComputerAgent - INFO - LLM processing started with 10 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 10 messages\n", - "\u001b[92m17:27:25 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/1909e6f5-b395-4e1d-b1f7-b06406f8731b/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:27:26 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 47%|██████████████████----------------------| 3447/7340 [121:08<136:48, 28.5 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 17:27:27,295 - agent.ComputerAgent - INFO - LLM processing started with 6 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 6 messages\n", - "\u001b[92m17:27:27 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m17:27:27 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/fde8bca8-8a90-4fed-b46f-c24829445665/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 17:27:27,985 - agent.ComputerAgent - INFO - Computer: click({'x': 262, 'y': 479})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 262, 'y': 479})\n", - " 47%|██████████████████----------------------| 3447/7340 [121:09<136:50, 28.4 steps/min]2025-08-11 17:27:28,643 - agent.ComputerAgent - INFO - LLM processing started with 18 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 18 messages\n", - "\u001b[92m17:27:28 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:27:29 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 47%|██████████████████----------------------| 3448/7340 [121:11<136:47, 28.5 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 17:27:30,354 - agent.ComputerAgent - INFO - LLM processing started with 32 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 32 messages\n", - "\u001b[92m17:27:30 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 47%|██████████████████----------------------| 3448/7340 [121:12<136:48, 28.4 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m17:27:30 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 17:27:31,027 - agent.ComputerAgent - INFO - Computer: click({'x': 219, 'y': 134})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 219, 'y': 134})\n", - " 47%|██████████████████----------------------| 3449/7340 [121:14<136:46, 28.4 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 17:27:33,880 - agent.ComputerAgent - INFO - Agent: Added the Dissolve slide transition to the first slide and saved the presentation.\n", - "\n", - "Task completed.\n", - "INFO:agent.ComputerAgent:Agent: Added the Dissolve slide transition to the first slide and saved the presentation.\n", - "\n", - "Task completed.\n", - "2025-08-11 17:27:34,565 - agent.ComputerAgent - INFO - Total usage:\n", - " - completion_tokens: 475\n", - " - prompt_tokens: 6691\n", - " - total_tokens: 7166\n", - " - completion_tokens_details:\n", - " - accepted_prediction_tokens: 0\n", - " - audio_tokens: 0\n", - " - reasoning_tokens: 448\n", - " - rejected_prediction_tokens: 0\n", - " - prompt_tokens_details:\n", - " - audio_tokens: 0\n", - " - cached_tokens: 4608\n", - " - response_cost: $0.0079\n", - "INFO:agent.ComputerAgent:Total usage:\n", - " - completion_tokens: 475\n", - " - prompt_tokens: 6691\n", - " - total_tokens: 7166\n", - " - completion_tokens_details:\n", - " - accepted_prediction_tokens: 0\n", - " - audio_tokens: 0\n", - " - reasoning_tokens: 448\n", - " - rejected_prediction_tokens: 0\n", - " - prompt_tokens_details:\n", - " - audio_tokens: 0\n", - " - cached_tokens: 4608\n", - " - response_cost: $0.0079\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - " 47%|██████████████████----------------------| 3450/7340 [121:17<136:45, 28.4 steps/min]\u001b[92m17:27:35 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/b6bbc5bc-5598-4043-be1e-6ebf2da5f046/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 17:27:35,955 - agent.ComputerAgent - INFO - LLM processing started with 38 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 38 messages\n", - "\u001b[92m17:27:35 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m17:27:35 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 17:27:36,603 - agent.ComputerAgent - INFO - Computer: move({'x': 166, 'y': 68})\n", - "INFO:agent.ComputerAgent:Computer: move({'x': 166, 'y': 68})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/4b18a76d-ef46-4622-9643-9ee6fe4900a3/invoke \"HTTP/1.1 200 OK\"\n", - " 47%|██████████████████----------------------| 3450/7340 [121:18<136:46, 28.4 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/b3b14802-9f99-46f5-8fa9-9661af7a973d/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 17:27:37,265 - agent.ComputerAgent - INFO - LLM processing started with 34 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 34 messages\n", - "\u001b[92m17:27:37 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m17:27:37 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 47%|██████████████████----------------------| 3451/7340 [121:19<136:43, 28.4 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m17:27:38 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 17:27:39,152 - agent.ComputerAgent - INFO - Computer: click({'x': 87, 'y': 158})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 87, 'y': 158})\n", - " 47%|██████████████████----------------------| 3451/7340 [121:20<136:44, 28.4 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:27:39 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 17:27:41,144 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'ctrl+f'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'ctrl+f'})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - " 47%|██████████████████----------------------| 3452/7340 [121:23<136:43, 28.4 steps/min]\u001b[92m17:27:41 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m17:27:41 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 17:27:42,465 - agent.ComputerAgent - INFO - LLM processing started with 34 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 34 messages\n", - "\u001b[92m17:27:42 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 17:27:43,111 - agent.ComputerAgent - INFO - Computer: double_click({'x': 984, 'y': 145})\n", - "INFO:agent.ComputerAgent:Computer: double_click({'x': 984, 'y': 145})\n", - "\u001b[92m17:27:43 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/57944bbf-74a1-4e6d-9401-f7b0144460f7/invoke \"HTTP/1.1 200 OK\"\n", - " 47%|██████████████████----------------------| 3452/7340 [121:24<136:45, 28.4 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m17:27:43 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 17:27:44,178 - agent.ComputerAgent - INFO - LLM processing started with 26 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 26 messages\n", - "\u001b[92m17:27:44 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:27:44 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 17:27:45,504 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'ctrl+a'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'ctrl+a'})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:27:46 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/b3b14802-9f99-46f5-8fa9-9661af7a973d/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 47%|██████████████████----------------------| 3453/7340 [121:27<136:43, 28.4 steps/min]2025-08-11 17:27:46,827 - agent.ComputerAgent - INFO - Computer: drag({'path': [{'x': 86, 'y': 123}, {'x': 83, 'y': 250}]})\n", - "INFO:agent.ComputerAgent:Computer: drag({'path': [{'x': 86, 'y': 123}, {'x': 83, 'y': 250}]})\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 17:27:47,486 - agent.ComputerAgent - INFO - LLM processing started with 26 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 26 messages\n", - "\u001b[92m17:27:47 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 47%|██████████████████----------------------| 3466/7340 [121:29<135:47, 28.5 steps/min]\u001b[92m17:27:47 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 17:27:48,140 - agent.ComputerAgent - INFO - Computer: click({'x': 225, 'y': 564})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 225, 'y': 564})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 17:27:49,491 - agent.ComputerAgent - INFO - Computer: type({'text': 'ls -1 *.doc\\n'})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': 'ls -1 *.doc\\n'})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v1/environments/a4b4d291-1fca-4038-8670-448014a55182/reset \"HTTP/1.1 200 OK\"\n", - " 47%|██████████████████----------------------| 3467/7340 [121:31<135:45, 28.5 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/b3b14802-9f99-46f5-8fa9-9661af7a973d/close \"HTTP/1.1 200 OK\"\n", - " 47%|██████████████████----------------------| 3469/7340 [121:32<135:38, 28.5 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/dc026dd3-8d59-43e0-a475-ecef72f1db12/invoke \"HTTP/1.1 200 OK\"\n", - " 47%|██████████████████----------------------| 3469/7340 [121:33<135:39, 28.5 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:27:52 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 17:27:53,445 - agent.ComputerAgent - INFO - LLM processing started with 30 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 30 messages\n", - "\u001b[92m17:27:53 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: GET https://orchestration.hud.so/hud-gym/api/v1/gyms/OSWorld-Ubuntu \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/a4b4d291-1fca-4038-8670-448014a55182/invoke \"HTTP/1.1 200 OK\"\n", - "Loading checkpoint shards: 0%| | 0/4 [00:00\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m17:28:59 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m17:29:00 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 17:29:01,218 - agent.ComputerAgent - INFO - Computer: click({'x': 521, 'y': 422})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 521, 'y': 422})\n", - "2025-08-11 17:29:01,891 - agent.ComputerAgent - INFO - Computer: click({'x': 17, 'y': 335})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 17, 'y': 335})\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 17:29:02,530 - agent.ComputerAgent - INFO - LLM processing started with 34 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 34 messages\n", - "\u001b[92m17:29:02 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 48%|███████████████████---------------------| 3490/7340 [122:44<135:23, 28.4 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m17:29:02 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m17:29:02 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 17:29:03,204 - agent.ComputerAgent - INFO - Agent: Taking a screenshot to see the current computer screen.\n", - "INFO:agent.ComputerAgent:Agent: Taking a screenshot to see the current computer screen.\n", - "2025-08-11 17:29:03,205 - agent.ComputerAgent - INFO - Computer: click({'button': 'left', 'x': 960, 'y': 713})\n", - "INFO:agent.ComputerAgent:Computer: click({'button': 'left', 'x': 960, 'y': 713})\n", - "2025-08-11 17:29:04,634 - agent.ComputerAgent - INFO - Computer: scroll({'scroll_y': -646, 'scroll_x': 0, 'x': 890, 'y': 760})\n", - "INFO:agent.ComputerAgent:Computer: scroll({'scroll_y': -646, 'scroll_x': 0, 'x': 890, 'y': 760})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - " 48%|███████████████████---------------------| 3492/7340 [122:47<135:18, 28.4 steps/min]\u001b[92m17:29:05 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 17:29:05,965 - agent.ComputerAgent - INFO - LLM processing started with 42 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 42 messages\n", - "\u001b[92m17:29:05 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m17:29:06 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 17:29:06,998 - agent.ComputerAgent - INFO - Agent: Taking a screenshot to see the current computer screen.\n", - "INFO:agent.ComputerAgent:Agent: Taking a screenshot to see the current computer screen.\n", - "2025-08-11 17:29:06,998 - agent.ComputerAgent - INFO - Computer: click({'x': 15, 'y': 432})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 15, 'y': 432})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 17:29:08,322 - agent.ComputerAgent - INFO - Computer: wait({})\n", - "INFO:agent.ComputerAgent:Computer: wait({})\n", - " 48%|███████████████████---------------------| 3496/7340 [122:54<135:08, 28.4 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/57944bbf-74a1-4e6d-9401-f7b0144460f7/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/3bde46c9-685b-4102-9ef4-a1535d5fcc85/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/479a3737-3ad4-48da-b73f-c8ea6e38d096/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 17:29:13,557 - agent.ComputerAgent - INFO - LLM processing started with 32 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 32 messages\n", - "\u001b[92m17:29:13 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/2a28af1e-e61d-489c-a18e-23c5071c9aff/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/a4b4d291-1fca-4038-8670-448014a55182/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/1909e6f5-b395-4e1d-b1f7-b06406f8731b/invoke \"HTTP/1.1 200 OK\"\n", - " 48%|███████████████████---------------------| 3496/7340 [122:55<135:09, 28.4 steps/min]2025-08-11 17:29:14,241 - agent.ComputerAgent - INFO - LLM processing started with 6 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 6 messages\n", - "\u001b[92m17:29:14 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/0322563b-daf3-41ae-8a08-f5ecd9282bcc/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/create_environment \"HTTP/1.1 200 OK\"\n", - "2025-08-11 17:29:14,878 - agent.ComputerAgent - INFO - LLM processing started with 6 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 6 messages\n", - "\u001b[92m17:29:14 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 17:29:15,558 - agent.ComputerAgent - INFO - LLM processing started with 10 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 10 messages\n", - "\u001b[92m17:29:15 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - " 48%|███████████████████---------------------| 3496/7340 [122:57<135:12, 28.4 steps/min]\u001b[92m17:29:16 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/792a6953-2092-47e4-a8a8-57a4af4e3be1/invoke \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 17:29:16,889 - agent.ComputerAgent - INFO - LLM processing started with 14 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 14 messages\n", - "\u001b[92m17:29:16 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m17:29:16 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 17:29:17,556 - agent.ComputerAgent - INFO - Computer: click({'x': 599, 'y': 760})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 599, 'y': 760})\n", - " 48%|███████████████████---------------------| 3496/7340 [122:59<135:13, 28.4 steps/min]2025-08-11 17:29:18,237 - agent.ComputerAgent - INFO - LLM processing started with 8 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 8 messages\n", - "\u001b[92m17:29:18 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 17:29:18,917 - agent.ComputerAgent - INFO - LLM processing started with 38 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 38 messages\n", - "\u001b[92m17:29:18 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 48%|███████████████████---------------------| 3497/7340 [123:00<135:10, 28.4 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 17:29:19,548 - agent.ComputerAgent - INFO - LLM processing started with 16 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 16 messages\n", - "\u001b[92m17:29:19 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/a5f69ad6-9361-4670-b101-61761113341c/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 48%|███████████████████---------------------| 3497/7340 [123:01<135:12, 28.4 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:29:21 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 48%|███████████████████---------------------| 3497/7340 [123:03<135:14, 28.4 steps/min]\u001b[92m17:29:21 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 17:29:23,251 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'enter'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'enter'})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:29:23 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m17:29:23 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m17:29:23 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/fde8bca8-8a90-4fed-b46f-c24829445665/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:29:24 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 48%|███████████████████---------------------| 3497/7340 [123:06<135:17, 28.4 steps/min]\u001b[92m17:29:25 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m17:29:25 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 17:29:25,847 - agent.ComputerAgent - INFO - Computer: scroll({'scroll_y': 612, 'x': 655, 'y': 419})\n", - "INFO:agent.ComputerAgent:Computer: scroll({'scroll_y': 612, 'x': 655, 'y': 419})\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m17:29:26 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m17:29:26 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - " 48%|███████████████████---------------------| 3499/7340 [123:08<135:11, 28.4 steps/min]\u001b[92m17:29:27 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 17:29:27,867 - agent.ComputerAgent - INFO - LLM processing started with 24 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 24 messages\n", - "\u001b[92m17:29:27 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 17:29:28,508 - agent.ComputerAgent - INFO - Computer: click({'x': 256, 'y': 128})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 256, 'y': 128})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:29:28 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:29:29 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:29:29 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 17:29:31,110 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'ctrl+a'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'ctrl+a'})\n", - "\u001b[92m17:29:31 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m17:29:31 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m17:29:31 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 17:29:31,743 - agent.ComputerAgent - INFO - Computer: click({'x': 182, 'y': 105})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 182, 'y': 105})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:29:31 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m17:29:32 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 48%|███████████████████---------------------| 3499/7340 [123:14<135:16, 28.4 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 17:29:33,138 - agent.ComputerAgent - INFO - Computer: click({'x': 634, 'y': 529})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 634, 'y': 529})\n", - "2025-08-11 17:29:33,791 - agent.ComputerAgent - INFO - Computer: drag({'path': [{'x': 82, 'y': 124}, {'x': 75, 'y': 124}]})\n", - "INFO:agent.ComputerAgent:Computer: drag({'path': [{'x': 82, 'y': 124}, {'x': 75, 'y': 124}]})\n", - "2025-08-11 17:29:34,450 - agent.ComputerAgent - INFO - Computer: scroll({'scroll_y': -517, 'scroll_x': 0, 'x': 46, 'y': 762})\n", - "INFO:agent.ComputerAgent:Computer: scroll({'scroll_y': -517, 'scroll_x': 0, 'x': 46, 'y': 762})\n", - "\u001b[92m17:29:34 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m17:29:34 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 17:29:35,110 - agent.ComputerAgent - INFO - LLM processing started with 34 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 34 messages\n", - "\u001b[92m17:29:35 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m17:29:35 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 17:29:35,790 - agent.ComputerAgent - INFO - Computer: scroll({'scroll_y': 673, 'scroll_x': 0, 'x': 86, 'y': 245})\n", - "INFO:agent.ComputerAgent:Computer: scroll({'scroll_y': 673, 'scroll_x': 0, 'x': 86, 'y': 245})\n", - "\u001b[92m17:29:35 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 48%|███████████████████---------------------| 3501/7340 [123:17<135:11, 28.4 steps/min]2025-08-11 17:29:36,477 - agent.ComputerAgent - INFO - Computer: click({'x': 15, 'y': 333})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 15, 'y': 333})\n", - "\u001b[92m17:29:36 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 17:29:37,144 - agent.ComputerAgent - INFO - Computer: click({'x': 268, 'y': 329})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 268, 'y': 329})\n", - "2025-08-11 17:29:37,807 - agent.ComputerAgent - INFO - Computer: drag({'path': [{'x': 332, 'y': 308}, {'x': 345, 'y': 308}]})\n", - "INFO:agent.ComputerAgent:Computer: drag({'path': [{'x': 332, 'y': 308}, {'x': 345, 'y': 308}]})\n", - " 48%|███████████████████---------------------| 3508/7340 [123:20<134:44, 28.4 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/f8984906-7392-4305-88fa-ae9a4808fa8d/invoke \"HTTP/1.1 200 OK\"\n" - ] - }, - { - "name": "stdout", - "output_type": "stream", - "text": [ - "No screenshot found, taking screenshot\n" - ] - }, - { - "name": "stderr", - "output_type": "stream", - "text": [ - "2025-08-11 17:29:39,468 - agent.ComputerAgent - INFO - LLM processing started with 39 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 39 messages\n", - "\u001b[92m17:29:39 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 17:29:40,852 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'ctrl+f'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'ctrl+f'})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 17:29:42,174 - agent.ComputerAgent - INFO - Computer: wait({})\n", - "INFO:agent.ComputerAgent:Computer: wait({})\n", - " 48%|███████████████████---------------------| 3508/7340 [123:23<134:47, 28.4 steps/min]2025-08-11 17:29:43,192 - agent.ComputerAgent - INFO - LLM processing started with 40 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 40 messages\n", - "\u001b[92m17:29:43 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/a4b4d291-1fca-4038-8670-448014a55182/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/b6bbc5bc-5598-4043-be1e-6ebf2da5f046/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/479a3737-3ad4-48da-b73f-c8ea6e38d096/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/0322563b-daf3-41ae-8a08-f5ecd9282bcc/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/dc026dd3-8d59-43e0-a475-ecef72f1db12/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/050a0934-63e8-46a0-8868-de32b28174ef/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/4b18a76d-ef46-4622-9643-9ee6fe4900a3/invoke \"HTTP/1.1 200 OK\"\n", - " 48%|███████████████████---------------------| 3509/7340 [123:25<134:44, 28.4 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/94463065-a78e-479a-b964-45ad23a48cbb/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 17:29:43,878 - agent.ComputerAgent - INFO - LLM processing started with 16 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 16 messages\n", - "\u001b[92m17:29:43 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 17:29:44,537 - agent.ComputerAgent - INFO - LLM processing started with 12 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 12 messages\n", - "\u001b[92m17:29:44 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 17:29:45,557 - agent.ComputerAgent - INFO - LLM processing started with 8 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 8 messages\n", - "\u001b[92m17:29:45 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/b28701c2-0fa4-4b07-bace-735fd2133893/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 400 Bad Request\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:29:46 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/2a28af1e-e61d-489c-a18e-23c5071c9aff/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 48%|███████████████████---------------------| 3510/7340 [123:28<134:43, 28.4 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 17:29:47,830 - agent.ComputerAgent - INFO - LLM processing started with 42 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 42 messages\n", - "\u001b[92m17:29:47 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 17:29:48,461 - agent.ComputerAgent - INFO - LLM processing started with 16 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 16 messages\n", - "\u001b[92m17:29:48 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m17:29:48 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 48%|███████████████████---------------------| 3510/7340 [123:30<134:45, 28.4 steps/min]2025-08-11 17:29:49,532 - agent.ComputerAgent - INFO - Computer: click({'x': 728, 'y': 179})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 728, 'y': 179})\n", - "2025-08-11 17:29:50,219 - agent.ComputerAgent - INFO - LLM processing started with 36 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 36 messages\n", - "\u001b[92m17:29:50 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:29:50 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 48%|███████████████████---------------------| 3510/7340 [123:32<134:48, 28.4 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:29:51 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m17:29:51 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 17:29:52,661 - agent.ComputerAgent - INFO - LLM processing started with 10 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 10 messages\n", - "\u001b[92m17:29:52 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 17:29:53,352 - agent.ComputerAgent - INFO - Computer: double_click({'x': 181, 'y': 105})\n", - "INFO:agent.ComputerAgent:Computer: double_click({'x': 181, 'y': 105})\n", - "\u001b[92m17:29:53 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:29:54 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/f8984906-7392-4305-88fa-ae9a4808fa8d/invoke \"HTTP/1.1 200 OK\"\n", - " 48%|███████████████████---------------------| 3511/7340 [123:35<134:47, 28.4 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 17:29:54,670 - agent.ComputerAgent - INFO - Computer: click({'x': 399, 'y': 541})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 399, 'y': 541})\n", - "2025-08-11 17:29:55,330 - agent.ComputerAgent - INFO - LLM processing started with 41 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 41 messages\n", - "\u001b[92m17:29:55 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/b6bbc5bc-5598-4043-be1e-6ebf2da5f046/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - " 48%|███████████████████---------------------| 3513/7340 [123:37<134:40, 28.4 steps/min]\u001b[92m17:29:56 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 17:29:57,015 - agent.ComputerAgent - INFO - LLM processing started with 26 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 26 messages\n", - "\u001b[92m17:29:57 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m17:29:57 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/3bde46c9-685b-4102-9ef4-a1535d5fcc85/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 48%|███████████████████---------------------| 3514/7340 [123:38<134:37, 28.4 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 17:29:57,710 - agent.ComputerAgent - INFO - Computer: click({'x': 525, 'y': 400})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 525, 'y': 400})\n", - "\u001b[92m17:29:57 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m17:29:58 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 17:29:59,392 - agent.ComputerAgent - INFO - LLM processing started with 8 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 8 messages\n", - "\u001b[92m17:29:59 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 17:30:00,039 - agent.ComputerAgent - INFO - Computer: click({'x': 1009, 'y': 101})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 1009, 'y': 101})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/792a6953-2092-47e4-a8a8-57a4af4e3be1/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 400 Bad Request\"\n", - " 48%|███████████████████---------------------| 3515/7340 [123:41<134:36, 28.4 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 17:30:00,720 - agent.ComputerAgent - INFO - LLM processing started with 30 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 30 messages\n", - "\u001b[92m17:30:00 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m17:30:00 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 17:30:01,758 - agent.ComputerAgent - INFO - Computer: scroll({'scroll_y': -657, 'scroll_x': 0, 'x': 988, 'y': 427})\n", - "INFO:agent.ComputerAgent:Computer: scroll({'scroll_y': -657, 'scroll_x': 0, 'x': 988, 'y': 427})\n", - " 48%|███████████████████---------------------| 3517/7340 [123:43<134:29, 28.4 steps/min]2025-08-11 17:30:02,459 - agent.ComputerAgent - INFO - LLM processing started with 18 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 18 messages\n", - "\u001b[92m17:30:02 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m17:30:03 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/b6bbc5bc-5598-4043-be1e-6ebf2da5f046/close \"HTTP/1.1 200 OK\"\n", - " 48%|███████████████████---------------------| 3518/7340 [123:44<134:26, 28.4 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m17:30:03 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 17:30:04,432 - agent.ComputerAgent - INFO - Computer: scroll({'scroll_y': 638, 'scroll_x': 0, 'x': 90, 'y': 244})\n", - "INFO:agent.ComputerAgent:Computer: scroll({'scroll_y': 638, 'scroll_x': 0, 'x': 90, 'y': 244})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/f8984906-7392-4305-88fa-ae9a4808fa8d/invoke \"HTTP/1.1 200 OK\"\n", - " 48%|███████████████████---------------------| 3518/7340 [123:46<134:27, 28.4 steps/min]2025-08-11 17:30:05,089 - agent.ComputerAgent - INFO - LLM processing started with 43 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 43 messages\n", - "\u001b[92m17:30:05 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 48%|███████████████████---------------------| 3519/7340 [123:47<134:24, 28.4 steps/min]INFO:httpx:HTTP Request: GET https://orchestration.hud.so/hud-gym/api/v1/gyms/OSWorld-Ubuntu \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/57944bbf-74a1-4e6d-9401-f7b0144460f7/invoke \"HTTP/1.1 200 OK\"\n", - " 48%|███████████████████---------------------| 3519/7340 [123:48<134:25, 28.4 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/a4b4d291-1fca-4038-8670-448014a55182/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/fde8bca8-8a90-4fed-b46f-c24829445665/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/479a3737-3ad4-48da-b73f-c8ea6e38d096/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 17:30:07,300 - agent.ComputerAgent - INFO - LLM processing started with 36 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 36 messages\n", - "\u001b[92m17:30:07 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 400 Bad Request\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m17:30:07 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 17:30:08,649 - agent.ComputerAgent - INFO - LLM processing started with 14 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 14 messages\n", - "\u001b[92m17:30:08 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 48%|███████████████████---------------------| 3520/7340 [123:50<134:23, 28.4 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 17:30:09,313 - agent.ComputerAgent - INFO - LLM processing started with 26 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 26 messages\n", - "\u001b[92m17:30:09 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "Loading checkpoint shards: 0%| | 0/4 [00:00&1 | sed -n '1,120p'\\n\"})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': \"ffmpeg -hide_banner -i video.mp4 2>&1 | sed -n '1,120p'\\n\"})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/83c40b56-f0bf-4b3a-97a5-8a1ae567e0a1/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/8d107e49-ae48-4b20-a0a1-7facc71e66f7/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m18:10:31 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 62%|████████████████████████----------------| 4522/7340 [164:12<102:20, 27.5 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 18:10:31,841 - agent.ComputerAgent - INFO - Computer: click({'x': 802, 'y': 437})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 802, 'y': 437})\n", - "2025-08-11 18:10:32,505 - agent.ComputerAgent - INFO - Computer: click({'x': 745, 'y': 540})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 745, 'y': 540})\n", - "2025-08-11 18:10:33,146 - agent.ComputerAgent - INFO - LLM processing started with 30 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 30 messages\n", - "\u001b[92m18:10:33 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/f096381e-eb5b-49dc-8943-c821405cce10/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/afdc88be-f209-412c-8905-25f3e8cbf43a/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/53e1a378-de8f-4a22-9dc0-27eef85d8356/invoke \"HTTP/1.1 200 OK\"\n", - " 62%|████████████████████████----------------| 4523/7340 [164:14<102:17, 27.5 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 18:10:33,803 - agent.ComputerAgent - INFO - LLM processing started with 30 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 30 messages\n", - "\u001b[92m18:10:33 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m18:10:33 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 18:10:34,476 - agent.ComputerAgent - INFO - LLM processing started with 36 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 36 messages\n", - "\u001b[92m18:10:34 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 18:10:35,141 - agent.ComputerAgent - INFO - Computer: click({'x': 205, 'y': 152})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 205, 'y': 152})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m18:10:35 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/91803c09-cf12-4c24-92ec-24bcf68c0897/invoke \"HTTP/1.1 200 OK\"\n", - " 62%|████████████████████████----------------| 4525/7340 [164:17<102:12, 27.5 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 18:10:36,437 - agent.ComputerAgent - INFO - LLM processing started with 16 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 16 messages\n", - "\u001b[92m18:10:36 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m18:10:36 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m18:10:37 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 18:10:37,794 - agent.ComputerAgent - INFO - Computer: click({'x': 1014, 'y': 31})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 1014, 'y': 31})\n", - " 62%|████████████████████████----------------| 4526/7340 [164:19<102:10, 27.5 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 18:10:38,466 - agent.ComputerAgent - INFO - LLM processing started with 12 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 12 messages\n", - "\u001b[92m18:10:38 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m18:10:38 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 18:10:39,125 - agent.ComputerAgent - INFO - Computer: click({'x': 399, 'y': 354})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 399, 'y': 354})\n", - " 62%|████████████████████████----------------| 4527/7340 [164:20<102:07, 27.5 steps/min]2025-08-11 18:10:39,795 - agent.ComputerAgent - INFO - LLM processing started with 1 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 1 messages\n", - "\u001b[92m18:10:39 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m18:10:40 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 62%|████████████████████████----------------| 4528/7340 [164:22<102:04, 27.5 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m18:10:41 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m18:10:41 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 18:10:42,145 - agent.ComputerAgent - INFO - Computer: click({'x': 437, 'y': 99})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 437, 'y': 99})\n", - "\u001b[92m18:10:42 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/f55f73a3-1816-4f61-8ec1-88f743cec333/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/b01cd4a6-3203-476b-8ece-c651b889f821/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - " 62%|████████████████████████----------------| 4528/7340 [164:24<102:06, 27.5 steps/min]\u001b[92m18:10:42 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 18:10:43,477 - agent.ComputerAgent - INFO - Computer: click({'x': 46, 'y': 52})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 46, 'y': 52})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/941d9ec3-7c28-40f6-b948-70db95115571/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/91803c09-cf12-4c24-92ec-24bcf68c0897/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/c83605a3-e62d-48d7-8568-f181d5627773/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/fcdab7d3-0448-49dd-b2db-f79a7c74a08b/invoke \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 18:10:44,127 - agent.ComputerAgent - INFO - LLM processing started with 18 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 18 messages\n", - "\u001b[92m18:10:44 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 18:10:44,765 - agent.ComputerAgent - INFO - LLM processing started with 22 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 22 messages\n", - "\u001b[92m18:10:44 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 18:10:45,424 - agent.ComputerAgent - INFO - LLM processing started with 30 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 30 messages\n", - "\u001b[92m18:10:45 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m18:10:45 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 62%|████████████████████████----------------| 4545/7340 [164:27<101:07, 27.6 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/f0aa6a3e-e61f-49b1-ade9-e8150e333596/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/cb48f65f-d00e-465a-a0ea-394e844382ca/invoke \"HTTP/1.1 200 OK\"\n", - "\u001b[92m18:10:45 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 18:10:46,114 - agent.ComputerAgent - INFO - LLM processing started with 38 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 38 messages\n", - "\u001b[92m18:10:46 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m18:10:46 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 18:10:46,753 - agent.ComputerAgent - INFO - Computer: drag({'path': [{'x': 51, 'y': 730}, {'x': 991, 'y': 759}]})\n", - "INFO:agent.ComputerAgent:Computer: drag({'path': [{'x': 51, 'y': 730}, {'x': 991, 'y': 759}]})\n", - "2025-08-11 18:10:47,399 - agent.ComputerAgent - INFO - LLM processing started with 24 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 24 messages\n", - "\u001b[92m18:10:47 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 62%|████████████████████████----------------| 4546/7340 [164:29<101:05, 27.6 steps/min]2025-08-11 18:10:48,065 - agent.ComputerAgent - INFO - LLM processing started with 1 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 1 messages\n", - "\u001b[92m18:10:48 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m18:10:48 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/91803c09-cf12-4c24-92ec-24bcf68c0897/close \"HTTP/1.1 200 OK\"\n", - " 62%|████████████████████████----------------| 4547/7340 [164:30<101:02, 27.6 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m18:10:49 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m18:10:50 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 18:10:50,780 - agent.ComputerAgent - INFO - Agent: Taking a screenshot to see the current computer screen.\n", - "INFO:agent.ComputerAgent:Agent: Taking a screenshot to see the current computer screen.\n", - "2025-08-11 18:10:50,781 - agent.ComputerAgent - INFO - Computer: click({'x': 18, 'y': 237})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 18, 'y': 237})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m18:10:51 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 62%|████████████████████████----------------| 4547/7340 [164:33<101:04, 27.6 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Loading checkpoint shards: 0%| | 0/4 [00:00\n", - " 64%|█████████████████████████---------------| 4702/7340 [170:06<95:26, 27.6 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/f55f73a3-1816-4f61-8ec1-88f743cec333/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/2bf9cd89-2d6a-4856-a09d-a771bc278600/invoke \"HTTP/1.1 200 OK\"\n", - "Loading checkpoint shards: 0%| | 0/4 [00:00 Settings > Privacy and security > Security > Safe Browsing.\n", - "\n", - "Task completed\n", - "INFO:agent.ComputerAgent:Agent: Chrome’s Safe Browsing is now set to Enhanced protection. You’ll get real-time warnings about potentially harmful or deceptive websites, dangerous downloads, and malicious extensions.\n", - "\n", - "If you ever want to change this: Chrome menu > Settings > Privacy and security > Security > Safe Browsing.\n", - "\n", - "Task completed\n", - "2025-08-11 18:22:14,920 - agent.ComputerAgent - INFO - Total usage:\n", - " - completion_tokens: 514\n", - " - prompt_tokens: 6726\n", - " - total_tokens: 7240\n", - " - completion_tokens_details:\n", - " - accepted_prediction_tokens: 0\n", - " - audio_tokens: 0\n", - " - reasoning_tokens: 448\n", - " - rejected_prediction_tokens: 0\n", - " - prompt_tokens_details:\n", - " - audio_tokens: 0\n", - " - cached_tokens: 0\n", - " - response_cost: $0.0135\n", - "INFO:agent.ComputerAgent:Total usage:\n", - " - completion_tokens: 514\n", - " - prompt_tokens: 6726\n", - " - total_tokens: 7240\n", - " - completion_tokens_details:\n", - " - accepted_prediction_tokens: 0\n", - " - audio_tokens: 0\n", - " - reasoning_tokens: 448\n", - " - rejected_prediction_tokens: 0\n", - " - prompt_tokens_details:\n", - " - audio_tokens: 0\n", - " - cached_tokens: 0\n", - " - response_cost: $0.0135\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m18:22:15 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/26d5566b-d949-4b71-accb-45197078f693/invoke \"HTTP/1.1 200 OK\"\n", - " 66%|██████████████████████████--------------| 4823/7340 [175:57<91:49, 27.4 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 18:22:16,260 - agent.ComputerAgent - INFO - LLM processing started with 21 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 21 messages\n", - "\u001b[92m18:22:16 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 66%|██████████████████████████--------------| 4823/7340 [175:58<91:50, 27.4 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/cfe4e097-0434-4025-a00a-78e26d753e51/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 400 Bad Request\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/d9bc2461-8bd1-4c45-bebd-f473293c581c/invoke \"HTTP/1.1 200 OK\"\n", - " 66%|██████████████████████████--------------| 4824/7340 [175:59<91:47, 27.4 steps/min]2025-08-11 18:22:17,920 - agent.ComputerAgent - INFO - LLM processing started with 20 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 20 messages\n", - "\u001b[92m18:22:17 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/26d5566b-d949-4b71-accb-45197078f693/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 18:22:19,103 - agent.ComputerAgent - INFO - LLM processing started with 23 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 23 messages\n", - " 66%|██████████████████████████--------------| 4824/7340 [176:00<91:48, 27.4 steps/min]\u001b[92m18:22:19 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m18:22:19 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 18:22:19,764 - agent.ComputerAgent - INFO - Computer: click({'x': 120, 'y': 53})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 120, 'y': 53})\n", - " 66%|██████████████████████████--------------| 4824/7340 [176:01<91:48, 27.4 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/d9bc2461-8bd1-4c45-bebd-f473293c581c/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 400 Bad Request\"\n", - " 66%|██████████████████████████--------------| 4838/7340 [176:02<91:02, 27.5 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/d9bc2461-8bd1-4c45-bebd-f473293c581c/close \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m18:22:21 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/26d5566b-d949-4b71-accb-45197078f693/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/a4a2a38e-bec8-46b5-b9c9-3e82144e6ff7/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 18:22:22,630 - agent.ComputerAgent - INFO - LLM processing started with 25 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 25 messages\n", - "\u001b[92m18:22:22 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 18:22:23,272 - agent.ComputerAgent - INFO - Computer: click({'x': 833, 'y': 385})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 833, 'y': 385})\n", - " 66%|██████████████████████████--------------| 4838/7340 [176:05<91:03, 27.5 steps/min]2025-08-11 18:22:23,933 - agent.ComputerAgent - INFO - LLM processing started with 10 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 10 messages\n", - "\u001b[92m18:22:23 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: GET https://orchestration.hud.so/hud-gym/api/v1/gyms/OSWorld-Ubuntu \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 18:22:25,259 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'ctrl+alt+t'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'ctrl+alt+t'})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 400 Bad Request\"\n", - " 66%|██████████████████████████--------------| 4840/7340 [176:07<90:58, 27.5 steps/min]\u001b[92m18:22:25 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 18:22:25,941 - agent.ComputerAgent - INFO - Computer: click({'x': 989, 'y': 73})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 989, 'y': 73})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/06215fad-881b-4e96-84a9-854f2d453fc5/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 18:22:26,601 - agent.ComputerAgent - INFO - LLM processing started with 20 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 20 messages\n", - "\u001b[92m18:22:26 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 66%|██████████████████████████--------------| 4840/7340 [176:08<90:58, 27.5 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 18:22:27,300 - agent.ComputerAgent - INFO - LLM processing started with 8 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 8 messages\n", - "\u001b[92m18:22:27 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m18:22:27 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 18:22:27,963 - agent.ComputerAgent - INFO - Computer: click({'x': 901, 'y': 312})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 901, 'y': 312})\n", - " 66%|██████████████████████████--------------| 4841/7340 [176:09<90:56, 27.5 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m18:22:28 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/26d5566b-d949-4b71-accb-45197078f693/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/ad87d89c-437d-4ed4-b0f0-a157e7d11bbd/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 18:22:29,233 - agent.ComputerAgent - INFO - LLM processing started with 27 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 27 messages\n", - "\u001b[92m18:22:29 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "Loading checkpoint shards: 0%| | 0/4 [00:00 Lock).\n", - "2) Select the text box from the Layers/Selection panel (don’t click on the canvas).\n", - "3) Snap it to the left edge:\n", - "- Either set the X position to 0.\n", - "- Or use the Align Left to canvas/page/artboard command.\n", - "4) Unlock/show the image layer again if needed.\n", - "\n", - "Exact steps by app\n", - "- Figma:\n", - " - Lock image: Ctrl+Shift+L (Cmd+Shift+L on Mac) or click the lock icon in Layers.\n", - " - Select the text layer in Layers.\n", - " - In the right panel, set X = 0. Or select the frame and the text, then click Align left. Tip: Cmd/Ctrl+Click cycles through stacked layers; Right‑click > Select layer also helps.\n", - "\n", - "- Photoshop:\n", - " - Lock image layer (padlock in Layers).\n", - " - Select text layer in Layers.\n", - " - Press V for Move tool. Press Ctrl+A (Cmd+A) to Select All, click Align left edges in the top options bar, then Ctrl+D (Cmd+D) to deselect. Alternatively, Window > Properties and set X = 0 px.\n", - "\n", - "- Illustrator:\n", - " - Lock image: Object > Lock > Selection (Ctrl+2 / Cmd+2) or lock in Layers.\n", - " - Select text object.\n", - " - Align panel: set Align To = Artboard, then Horizontal Align Left. Or Transform panel X = 0.\n", - "\n", - "- InDesign:\n", - " - Lock image (Ctrl+L / Cmd+L) or via Layers.\n", - " - Select text frame.\n", - " - Window > Object & Layout > Align. Set Align To = Page, then Align Left. Or set X = 0 in the Control/Properties panel.\n", - "\n", - "- Canva:\n", - " - Select image > Lock (padlock on toolbar).\n", - " - Select text > Position > Left. Nudge with arrow keys if needed; it will snap to the page edge.\n", - "\n", - "- PowerPoint:\n", - " - Home > Select > Selection Pane: click the eye to hide the image (or select the text from the pane).\n", - " - Shape Format > Align > Align to Slide > Align Left.\n", - "\n", - "- Google Slides:\n", - " - Arrange > Order to move im...\n", - "INFO:agent.ComputerAgent:Agent: I can’t see your design app/document open on this computer, so I can’t move the text box directly. Here’s the fastest, foolproof way to get your text box flush left without ever clicking the image again:\n", - "\n", - "Universal quick steps (work in any editor)\n", - "1) Lock or hide the image layer.\n", - "- Use the Layers/Selection panel and click the lock icon on the image layer (or right‑click > Lock).\n", - "2) Select the text box from the Layers/Selection panel (don’t click on the canvas).\n", - "3) Snap it to the left edge:\n", - "- Either set the X position to 0.\n", - "- Or use the Align Left to canvas/page/artboard command.\n", - "4) Unlock/show the image layer again if needed.\n", - "\n", - "Exact steps by app\n", - "- Figma:\n", - " - Lock image: Ctrl+Shift+L (Cmd+Shift+L on Mac) or click the lock icon in Layers.\n", - " - Select the text layer in Layers.\n", - " - In the right panel, set X = 0. Or select the frame and the text, then click Align left. Tip: Cmd/Ctrl+Click cycles through stacked layers; Right‑click > Select layer also helps.\n", - "\n", - "- Photoshop:\n", - " - Lock image layer (padlock in Layers).\n", - " - Select text layer in Layers.\n", - " - Press V for Move tool. Press Ctrl+A (Cmd+A) to Select All, click Align left edges in the top options bar, then Ctrl+D (Cmd+D) to deselect. Alternatively, Window > Properties and set X = 0 px.\n", - "\n", - "- Illustrator:\n", - " - Lock image: Object > Lock > Selection (Ctrl+2 / Cmd+2) or lock in Layers.\n", - " - Select text object.\n", - " - Align panel: set Align To = Artboard, then Horizontal Align Left. Or Transform panel X = 0.\n", - "\n", - "- InDesign:\n", - " - Lock image (Ctrl+L / Cmd+L) or via Layers.\n", - " - Select text frame.\n", - " - Window > Object & Layout > Align. Set Align To = Page, then Align Left. Or set X = 0 in the Control/Properties panel.\n", - "\n", - "- Canva:\n", - " - Select image > Lock (padlock on toolbar).\n", - " - Select text > Position > Left. Nudge with arrow keys if needed; it will snap to the page edge.\n", - "\n", - "- PowerPoint:\n", - " - Home > Select > Selection Pane: click the eye to hide the image (or select the text from the pane).\n", - " - Shape Format > Align > Align to Slide > Align Left.\n", - "\n", - "- Google Slides:\n", - " - Arrange > Order to move im...\n", - "2025-08-11 18:35:23,380 - agent.ComputerAgent - INFO - Total usage:\n", - " - completion_tokens: 3277\n", - " - prompt_tokens: 1951\n", - " - total_tokens: 5228\n", - " - completion_tokens_details:\n", - " - accepted_prediction_tokens: 0\n", - " - audio_tokens: 0\n", - " - reasoning_tokens: 2624\n", - " - rejected_prediction_tokens: 0\n", - " - prompt_tokens_details:\n", - " - audio_tokens: 0\n", - " - cached_tokens: 0\n", - " - response_cost: $0.0352\n", - "INFO:agent.ComputerAgent:Total usage:\n", - " - completion_tokens: 3277\n", - " - prompt_tokens: 1951\n", - " - total_tokens: 5228\n", - " - completion_tokens_details:\n", - " - accepted_prediction_tokens: 0\n", - " - audio_tokens: 0\n", - " - reasoning_tokens: 2624\n", - " - rejected_prediction_tokens: 0\n", - " - prompt_tokens_details:\n", - " - audio_tokens: 0\n", - " - cached_tokens: 0\n", - " - response_cost: $0.0352\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/d05e9e78-ad03-41fc-a347-043ec46bd299/invoke \"HTTP/1.1 200 OK\"\n", - " 70%|████████████████████████████------------| 5147/7340 [189:05<80:33, 27.2 steps/min]2025-08-11 18:35:24,059 - agent.ComputerAgent - INFO - LLM processing started with 33 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 33 messages\n", - "\u001b[92m18:35:24 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 70%|████████████████████████████------------| 5147/7340 [189:06<80:34, 27.2 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m18:35:25 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - " 70%|████████████████████████████------------| 5147/7340 [189:07<80:34, 27.2 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/c268b680-eafe-4b8d-914a-28e5540231cd/invoke \"HTTP/1.1 200 OK\"\n", - "\u001b[92m18:35:25 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m18:35:25 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m18:35:26 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/0d923fcd-4666-4869-8ad2-17460c904167/invoke \"HTTP/1.1 200 OK\"\n", - "\u001b[92m18:35:26 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 70%|████████████████████████████------------| 5147/7340 [189:08<80:35, 27.2 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 18:35:27,111 - agent.ComputerAgent - INFO - Computer: drag({'path': [{'x': 76, 'y': 181}, {'x': 124, 'y': 181}]})\n", - "INFO:agent.ComputerAgent:Computer: drag({'path': [{'x': 76, 'y': 181}, {'x': 124, 'y': 181}]})\n", - "\u001b[92m18:35:27 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 18:35:27,806 - agent.ComputerAgent - INFO - Computer: click({'x': 290, 'y': 149})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 290, 'y': 149})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 400 Bad Request\"\n", - " 70%|████████████████████████████------------| 5148/7340 [189:09<80:32, 27.2 steps/min]2025-08-11 18:35:28,462 - agent.ComputerAgent - INFO - LLM processing started with 32 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 32 messages\n", - "\u001b[92m18:35:28 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 70%|████████████████████████████------------| 5150/7340 [189:10<80:26, 27.2 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/d05e9e78-ad03-41fc-a347-043ec46bd299/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 18:35:29,639 - agent.ComputerAgent - INFO - LLM processing started with 35 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 35 messages\n", - "\u001b[92m18:35:29 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 70%|████████████████████████████------------| 5150/7340 [189:13<80:28, 27.2 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 400 Bad Request\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/2350cfc0-8e2e-4302-b87c-df7d35bd6eea/invoke \"HTTP/1.1 200 OK\"\n", - " 70%|████████████████████████████------------| 5151/7340 [189:14<80:25, 27.2 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/d05e9e78-ad03-41fc-a347-043ec46bd299/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 18:35:33,840 - agent.ComputerAgent - INFO - LLM processing started with 37 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 37 messages\n", - "\u001b[92m18:35:33 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/89880137-9134-4973-9389-b3535802254c/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 70%|████████████████████████████------------| 5151/7340 [189:15<80:25, 27.2 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m18:35:34 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - " 70%|████████████████████████████------------| 5151/7340 [189:16<80:26, 27.2 steps/min]\u001b[92m18:35:35 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 18:35:35,720 - agent.ComputerAgent - INFO - Computer: click({'x': 207, 'y': 488})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 207, 'y': 488})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 400 Bad Request\"\n", - " 70%|████████████████████████████------------| 5152/7340 [189:17<80:23, 27.2 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/2350cfc0-8e2e-4302-b87c-df7d35bd6eea/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/c268b680-eafe-4b8d-914a-28e5540231cd/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 18:35:38,050 - agent.ComputerAgent - INFO - Agent: I converted the opened ODS on your Desktop to a CSV with the same base name while LibreOffice was running.\n", - "\n", - "Command used:\n", - "mkdir -p /tmp/lo-temp-profile && libreoffice --headless -env:UserInstallation=file:///tmp/lo-temp-profile --convert-to \"csv:Text - txt - csv (StarCalc):44,34,0\" --outdir ~/Desktop ~/Desktop/file_example_ODS_5000.ods\n", - "\n", - "Notes:\n", - "- A temporary LibreOffice profile was used (-env:UserInstallation=…) so the conversion works even with the GUI instance open.\n", - "- The CSV file has been created on your Desktop as file_example_ODS_5000.csv.\n", - "\n", - "Task completed.\n", - "INFO:agent.ComputerAgent:Agent: I converted the opened ODS on your Desktop to a CSV with the same base name while LibreOffice was running.\n", - "\n", - "Command used:\n", - "mkdir -p /tmp/lo-temp-profile && libreoffice --headless -env:UserInstallation=file:///tmp/lo-temp-profile --convert-to \"csv:Text - txt - csv (StarCalc):44,34,0\" --outdir ~/Desktop ~/Desktop/file_example_ODS_5000.ods\n", - "\n", - "Notes:\n", - "- A temporary LibreOffice profile was used (-env:UserInstallation=…) so the conversion works even with the GUI instance open.\n", - "- The CSV file has been created on your Desktop as file_example_ODS_5000.csv.\n", - "\n", - "Task completed.\n", - "2025-08-11 18:35:38,700 - agent.ComputerAgent - INFO - Total usage:\n", - " - completion_tokens: 920\n", - " - prompt_tokens: 10826\n", - " - total_tokens: 11746\n", - " - completion_tokens_details:\n", - " - accepted_prediction_tokens: 0\n", - " - audio_tokens: 0\n", - " - reasoning_tokens: 768\n", - " - rejected_prediction_tokens: 0\n", - " - prompt_tokens_details:\n", - " - audio_tokens: 0\n", - " - cached_tokens: 0\n", - " - response_cost: $0.0227\n", - "INFO:agent.ComputerAgent:Total usage:\n", - " - completion_tokens: 920\n", - " - prompt_tokens: 10826\n", - " - total_tokens: 11746\n", - " - completion_tokens_details:\n", - " - accepted_prediction_tokens: 0\n", - " - audio_tokens: 0\n", - " - reasoning_tokens: 768\n", - " - rejected_prediction_tokens: 0\n", - " - prompt_tokens_details:\n", - " - audio_tokens: 0\n", - " - cached_tokens: 0\n", - " - response_cost: $0.0227\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/2350cfc0-8e2e-4302-b87c-df7d35bd6eea/close \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/c268b680-eafe-4b8d-914a-28e5540231cd/close \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/d05e9e78-ad03-41fc-a347-043ec46bd299/invoke \"HTTP/1.1 200 OK\"\n", - " 70%|████████████████████████████------------| 5173/7340 [189:20<79:18, 27.3 steps/min]2025-08-11 18:35:39,393 - agent.ComputerAgent - INFO - LLM processing started with 39 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 39 messages\n", - "\u001b[92m18:35:39 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/055914cd-07b0-4dcd-9407-c6975b1eccbf/invoke \"HTTP/1.1 200 OK\"\n", - " 70%|████████████████████████████------------| 5173/7340 [189:22<79:19, 27.3 steps/min]2025-08-11 18:35:41,381 - agent.ComputerAgent - INFO - LLM processing started with 30 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 30 messages\n", - "\u001b[92m18:35:41 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: GET https://orchestration.hud.so/hud-gym/api/v1/gyms/OSWorld-Ubuntu \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v1/environments/655a0f34-fb5e-49f8-9a65-531af668d6c6/reset \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: GET https://orchestration.hud.so/hud-gym/api/v1/gyms/OSWorld-Ubuntu \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m18:35:42 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/5afdf327-0d8f-4749-8016-19cb1aedf273/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v1/environments/7a6ead00-3730-4f34-9acb-3c8109ec140a/reset \"HTTP/1.1 200 OK\"\n", - " 70%|████████████████████████████------------| 5173/7340 [189:24<79:20, 27.3 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 400 Bad Request\"\n", - " 70%|████████████████████████████------------| 5174/7340 [189:25<79:18, 27.3 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/5afdf327-0d8f-4749-8016-19cb1aedf273/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/d05e9e78-ad03-41fc-a347-043ec46bd299/invoke \"HTTP/1.1 200 OK\"\n", - "Loading checkpoint shards: 25%|██▌ | 1/4 [00:01<00:04, 1.67s/it]2025-08-11 18:35:45,241 - agent.ComputerAgent - INFO - LLM processing started with 41 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 41 messages\n", - "\u001b[92m18:35:45 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 71%|████████████████████████████------------| 5182/7340 [189:26<78:53, 27.4 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/5afdf327-0d8f-4749-8016-19cb1aedf273/close \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/7a6ead00-3730-4f34-9acb-3c8109ec140a/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/655a0f34-fb5e-49f8-9a65-531af668d6c6/invoke \"HTTP/1.1 200 OK\"\n", - "Loading checkpoint shards: 50%|█████ | 2/4 [00:03<00:03, 1.62s/it]27.4 steps/min]2025-08-11 18:35:47,491 - agent.ComputerAgent - INFO - LLM processing started with 1 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 1 messages\n", - "\u001b[92m18:35:47 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 400 Bad Request\"\n", - "INFO:httpx:HTTP Request: GET https://orchestration.hud.so/hud-gym/api/v1/gyms/OSWorld-Ubuntu \"HTTP/1.1 200 OK\"\n", - "Loading checkpoint shards: 75%|███████▌ | 3/4 [00:04<00:01, 1.58s/it]27.4 steps/min]2025-08-11 18:35:48,319 - agent.ComputerAgent - INFO - LLM processing started with 1 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 1 messages\n", - "\u001b[92m18:35:48 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "Loading checkpoint shards: 100%|██████████| 4/4 [00:05<00:00, 1.33s/it]27.4 steps/min]\n", - " 71%|████████████████████████████------------| 5183/7340 [189:31<78:52, 27.3 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/d05e9e78-ad03-41fc-a347-043ec46bd299/invoke \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 18:35:50,250 - agent.ComputerAgent - INFO - LLM processing started with 43 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 43 messages\n", - "\u001b[92m18:35:50 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 71%|████████████████████████████------------| 5183/7340 [189:32<78:52, 27.3 steps/min]\u001b[92m18:35:50 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 18:35:50,943 - agent.ComputerAgent - INFO - Computer: scroll({'scroll_y': 713, 'scroll_x': 0, 'x': 716, 'y': 646})\n", - "INFO:agent.ComputerAgent:Computer: scroll({'scroll_y': 713, 'scroll_x': 0, 'x': 716, 'y': 646})\n", - " 71%|████████████████████████████------------| 5184/7340 [189:34<78:50, 27.3 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 400 Bad Request\"\n", - " 71%|████████████████████████████------------| 5185/7340 [189:35<78:47, 27.3 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/d05e9e78-ad03-41fc-a347-043ec46bd299/invoke \"HTTP/1.1 200 OK\"\n", - " 71%|████████████████████████████------------| 5185/7340 [189:38<78:49, 27.3 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m18:35:57 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 71%|████████████████████████████------------| 5185/7340 [189:39<78:49, 27.3 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/9b006d7b-b853-41ed-8a84-b7eaa5b6e94b/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/d05e9e78-ad03-41fc-a347-043ec46bd299/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Loading checkpoint shards: 25%|██▌ | 1/4 [00:01<00:04, 1.66s/it]2025-08-11 18:35:59,836 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'ctrl+shift+p'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'ctrl+shift+p'})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/d05e9e78-ad03-41fc-a347-043ec46bd299/close \"HTTP/1.1 200 OK\"\n", - "Loading checkpoint shards: 50%|█████ | 2/4 [00:03<00:03, 1.61s/it]27.3 steps/min]2025-08-11 18:36:01,386 - agent.ComputerAgent - INFO - LLM processing started with 34 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 34 messages\n", - "\u001b[92m18:36:01 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 71%|████████████████████████████------------| 5185/7340 [189:43<78:51, 27.3 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 18:36:02,052 - agent.ComputerAgent - INFO - LLM processing started with 36 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 36 messages\n", - "\u001b[92m18:36:02 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "Loading checkpoint shards: 100%|██████████| 4/4 [00:05<00:00, 1.40s/it]27.3 steps/min]\n", - "INFO:httpx:HTTP Request: GET https://orchestration.hud.so/hud-gym/api/v1/gyms/OSWorld-Ubuntu \"HTTP/1.1 200 OK\"\n", - " 71%|████████████████████████████------------| 5185/7340 [189:46<78:52, 27.3 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m18:36:06 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 71%|████████████████████████████------------| 5185/7340 [189:47<78:52, 27.3 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Loading checkpoint shards: 25%|██▌ | 1/4 [00:01<00:04, 1.64s/it]27.3 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/89880137-9134-4973-9389-b3535802254c/invoke \"HTTP/1.1 200 OK\"\n", - " 71%|████████████████████████████------------| 5186/7340 [189:49<78:50, 27.3 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/create_environment \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/89880137-9134-4973-9389-b3535802254c/close \"HTTP/1.1 200 OK\"\n", - " 71%|████████████████████████████------------| 5186/7340 [189:51<78:51, 27.3 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/497d5104-1e6e-44a9-a164-fec745a337b6/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: GET https://orchestration.hud.so/hud-gym/api/v1/gyms/OSWorld-Ubuntu \"HTTP/1.1 200 OK\"\n", - "Loading checkpoint shards: 100%|██████████| 4/4 [00:05<00:00, 1.35s/it]27.3 steps/min]\n", - "\u001b[92m18:36:11 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 18:36:12,164 - agent.ComputerAgent - INFO - Agent: Taking a screenshot to see the current computer screen.\n", - "INFO:agent.ComputerAgent:Agent: Taking a screenshot to see the current computer screen.\n", - "2025-08-11 18:36:12,165 - agent.ComputerAgent - INFO - Computer: click({'x': 16, 'y': 428})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 16, 'y': 428})\n", - " 71%|████████████████████████████------------| 5186/7340 [189:54<78:52, 27.3 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - " 71%|████████████████████████████------------| 5187/7340 [189:56<78:50, 27.3 steps/min]\u001b[92m18:36:14 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 18:36:15,365 - agent.ComputerAgent - INFO - Agent: Taking a screenshot to see the current computer screen.\n", - "INFO:agent.ComputerAgent:Agent: Taking a screenshot to see the current computer screen.\n", - "2025-08-11 18:36:15,367 - agent.ComputerAgent - INFO - Computer: double_click({'x': 989, 'y': 650})\n", - "INFO:agent.ComputerAgent:Computer: double_click({'x': 989, 'y': 650})\n", - " 71%|████████████████████████████------------| 5188/7340 [189:59<78:48, 27.3 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/655a0f34-fb5e-49f8-9a65-531af668d6c6/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 18:36:18,079 - agent.ComputerAgent - INFO - LLM processing started with 6 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 6 messages\n", - "\u001b[92m18:36:18 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 71%|████████████████████████████------------| 5188/7340 [190:02<78:49, 27.3 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/7a6ead00-3730-4f34-9acb-3c8109ec140a/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 18:36:21,321 - agent.ComputerAgent - INFO - LLM processing started with 6 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 6 messages\n", - "\u001b[92m18:36:21 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 71%|████████████████████████████------------| 5188/7340 [190:04<78:50, 27.3 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 18:36:24,203 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'ctrl+c'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'ctrl+c'})\n", - " 71%|████████████████████████████------------| 5188/7340 [190:05<78:51, 27.3 steps/min]2025-08-11 18:36:25,341 - agent.ComputerAgent - INFO - LLM processing started with 32 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 32 messages\n", - "\u001b[92m18:36:25 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 71%|████████████████████████████------------| 5188/7340 [190:07<78:51, 27.3 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 71%|████████████████████████████------------| 5188/7340 [190:10<78:52, 27.3 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 18:36:29,780 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'pagedown'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'pagedown'})\n", - " 71%|████████████████████████████------------| 5189/7340 [190:12<78:50, 27.3 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m18:36:32 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 71%|████████████████████████████------------| 5189/7340 [190:13<78:51, 27.3 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Loading checkpoint shards: 50%|█████ | 2/4 [00:03<00:03, 1.58s/it]27.3 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/9b006d7b-b853-41ed-8a84-b7eaa5b6e94b/invoke \"HTTP/1.1 200 OK\"\n", - " 71%|████████████████████████████------------| 5189/7340 [190:17<78:53, 27.3 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Loading checkpoint shards: 75%|███████▌ | 3/4 [00:04<00:01, 1.56s/it]\u001b[92m18:36:37 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Loading checkpoint shards: 100%|██████████| 4/4 [00:05<00:00, 1.31s/it]27.3 steps/min]\n", - "2025-08-11 18:36:37,871 - agent.ComputerAgent - INFO - LLM processing started with 38 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 38 messages\n", - "\u001b[92m18:36:37 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 18:36:39,406 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'ctrl+shift+p'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'ctrl+shift+p'})\n", - " 71%|████████████████████████████------------| 5189/7340 [190:21<78:54, 27.3 steps/min]\u001b[92m18:36:39 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 18:36:40,047 - agent.ComputerAgent - INFO - Computer: click({'x': 15, 'y': 142})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 15, 'y': 142})\n", - "\u001b[92m18:36:40 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 18:36:40,733 - agent.ComputerAgent - INFO - Computer: click({'x': 15, 'y': 629})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 15, 'y': 629})\n", - " 71%|████████████████████████████------------| 5189/7340 [190:22<78:54, 27.3 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/0d923fcd-4666-4869-8ad2-17460c904167/invoke \"HTTP/1.1 200 OK\"\n", - " 71%|████████████████████████████------------| 5201/7340 [190:23<78:18, 27.3 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/0d923fcd-4666-4869-8ad2-17460c904167/close \"HTTP/1.1 200 OK\"\n", - " 71%|████████████████████████████------------| 5201/7340 [190:24<78:18, 27.3 steps/min]INFO:httpx:HTTP Request: GET https://orchestration.hud.so/hud-gym/api/v1/gyms/OSWorld-Ubuntu \"HTTP/1.1 200 OK\"\n", - " 71%|████████████████████████████------------| 5201/7340 [190:26<78:19, 27.3 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/7a6ead00-3730-4f34-9acb-3c8109ec140a/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 18:36:46,172 - agent.ComputerAgent - INFO - LLM processing started with 8 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 8 messages\n", - "\u001b[92m18:36:46 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m18:36:46 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/055914cd-07b0-4dcd-9407-c6975b1eccbf/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 71%|████████████████████████████------------| 5201/7340 [190:28<78:20, 27.3 steps/min]2025-08-11 18:36:47,482 - agent.ComputerAgent - INFO - LLM processing started with 34 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 34 messages\n", - "\u001b[92m18:36:47 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/create_environment \"HTTP/1.1 200 OK\"\n", - " 71%|████████████████████████████------------| 5201/7340 [190:30<78:21, 27.3 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/b3df65c5-9d1c-44fd-b9bb-37f1f0cd64dc/invoke \"HTTP/1.1 200 OK\"\n", - " 71%|████████████████████████████------------| 5201/7340 [190:32<78:21, 27.3 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/create_environment \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/create_environment \"HTTP/1.1 200 OK\"\n", - " 71%|████████████████████████████------------| 5201/7340 [190:33<78:22, 27.3 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/e42da596-e101-4fd3-9dea-8a1d63615dad/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/1fef1c7a-93ef-4a63-b067-399dfc4ff08a/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v1/environments/b3df65c5-9d1c-44fd-b9bb-37f1f0cd64dc/reset \"HTTP/1.1 200 OK\"\n", - "Loading checkpoint shards: 75%|███████▌ | 3/4 [00:04<00:01, 1.57s/it]27.3 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v1/environments/497d5104-1e6e-44a9-a164-fec745a337b6/reset \"HTTP/1.1 200 OK\"\n", - "Loading checkpoint shards: 100%|██████████| 4/4 [00:05<00:00, 1.32s/it]\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 18:36:55,357 - agent.ComputerAgent - INFO - Computer: wait({})\n", - "INFO:agent.ComputerAgent:Computer: wait({})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/b3df65c5-9d1c-44fd-b9bb-37f1f0cd64dc/invoke \"HTTP/1.1 200 OK\"\n", - " 71%|████████████████████████████------------| 5201/7340 [190:37<78:23, 27.3 steps/min]2025-08-11 18:36:56,039 - agent.ComputerAgent - INFO - LLM processing started with 1 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 1 messages\n", - "\u001b[92m18:36:56 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m18:36:56 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 18:36:56,709 - agent.ComputerAgent - INFO - Computer: click({'x': 15, 'y': 387})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 15, 'y': 387})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/497d5104-1e6e-44a9-a164-fec745a337b6/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v1/environments/e42da596-e101-4fd3-9dea-8a1d63615dad/reset \"HTTP/1.1 200 OK\"\n", - " 71%|████████████████████████████------------| 5202/7340 [190:38<78:21, 27.3 steps/min]2025-08-11 18:36:57,382 - agent.ComputerAgent - INFO - LLM processing started with 1 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 1 messages\n", - "\u001b[92m18:36:57 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v1/environments/1fef1c7a-93ef-4a63-b067-399dfc4ff08a/reset \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m18:36:58 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 71%|████████████████████████████------------| 5203/7340 [190:39<78:18, 27.3 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/e42da596-e101-4fd3-9dea-8a1d63615dad/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 18:36:58,641 - agent.ComputerAgent - INFO - LLM processing started with 1 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 1 messages\n", - "\u001b[92m18:36:58 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m18:36:58 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 18:36:59,325 - agent.ComputerAgent - INFO - Computer: click({'x': 368, 'y': 561})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 368, 'y': 561})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/create_environment \"HTTP/1.1 200 OK\"\n", - " 71%|████████████████████████████------------| 5203/7340 [190:41<78:19, 27.3 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/1fef1c7a-93ef-4a63-b067-399dfc4ff08a/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 18:37:00,501 - agent.ComputerAgent - INFO - LLM processing started with 1 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 1 messages\n", - "\u001b[92m18:37:00 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 71%|████████████████████████████------------| 5204/7340 [190:42<78:16, 27.3 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/af5d4b08-d761-4bf8-a2c7-7ce16ed98ab9/invoke \"HTTP/1.1 200 OK\"\n", - " 71%|████████████████████████████------------| 5204/7340 [190:43<78:16, 27.3 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m18:37:02 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/7a6ead00-3730-4f34-9acb-3c8109ec140a/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 18:37:03,933 - agent.ComputerAgent - INFO - Agent: Taking a screenshot to see the current computer screen.\n", - "INFO:agent.ComputerAgent:Agent: Taking a screenshot to see the current computer screen.\n", - "2025-08-11 18:37:03,935 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'meta'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'meta'})\n", - " 71%|████████████████████████████------------| 5204/7340 [190:45<78:17, 27.3 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m18:37:04 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 18:37:05,291 - agent.ComputerAgent - INFO - LLM processing started with 10 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 10 messages\n", - "\u001b[92m18:37:05 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/655a0f34-fb5e-49f8-9a65-531af668d6c6/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m18:37:05 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 18:37:06,012 - agent.ComputerAgent - INFO - Computer: click({'x': 524, 'y': 503})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 524, 'y': 503})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/055914cd-07b0-4dcd-9407-c6975b1eccbf/invoke \"HTTP/1.1 200 OK\"\n", - " 71%|████████████████████████████------------| 5205/7340 [190:47<78:15, 27.3 steps/min]\u001b[92m18:37:06 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 18:37:06,653 - agent.ComputerAgent - INFO - Agent: Taking a screenshot to see the current computer screen.\n", - "INFO:agent.ComputerAgent:Agent: Taking a screenshot to see the current computer screen.\n", - "2025-08-11 18:37:06,654 - agent.ComputerAgent - INFO - Computer: click({'x': 13, 'y': 41})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 13, 'y': 41})\n", - "2025-08-11 18:37:07,313 - agent.ComputerAgent - INFO - LLM processing started with 8 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 8 messages\n", - "\u001b[92m18:37:07 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 18:37:08,648 - agent.ComputerAgent - INFO - Agent: Taking a screenshot to see the current computer screen.\n", - "INFO:agent.ComputerAgent:Agent: Taking a screenshot to see the current computer screen.\n", - "2025-08-11 18:37:08,649 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'win'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'win'})\n", - " 71%|████████████████████████████------------| 5206/7340 [190:50<78:13, 27.3 steps/min]2025-08-11 18:37:09,323 - agent.ComputerAgent - INFO - LLM processing started with 36 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 36 messages\n", - "\u001b[92m18:37:09 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 71%|████████████████████████████------------| 5208/7340 [190:55<78:09, 27.3 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/9b006d7b-b853-41ed-8a84-b7eaa5b6e94b/invoke \"HTTP/1.1 200 OK\"\n", - " 71%|████████████████████████████------------| 5208/7340 [190:56<78:09, 27.3 steps/min]2025-08-11 18:37:15,613 - agent.ComputerAgent - INFO - LLM processing started with 40 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 40 messages\n", - "\u001b[92m18:37:15 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/b3df65c5-9d1c-44fd-b9bb-37f1f0cd64dc/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/e42da596-e101-4fd3-9dea-8a1d63615dad/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/1fef1c7a-93ef-4a63-b067-399dfc4ff08a/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:openai._base_client:Retrying request to /chat/completions in 0.403242 seconds\n", - " 71%|████████████████████████████------------| 5208/7340 [190:57<78:10, 27.3 steps/min]2025-08-11 18:37:16,773 - agent.ComputerAgent - INFO - LLM processing started with 6 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 6 messages\n", - "\u001b[92m18:37:16 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 18:37:18,458 - agent.ComputerAgent - INFO - Computer: type({'text': 'drive.google.com'})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': 'drive.google.com'})\n", - " 71%|████████████████████████████------------| 5208/7340 [191:00<78:11, 27.3 steps/min]2025-08-11 18:37:19,460 - agent.ComputerAgent - INFO - LLM processing started with 6 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 6 messages\n", - "\u001b[92m18:37:19 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 71%|████████████████████████████------------| 5209/7340 [191:01<78:08, 27.3 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m18:37:20 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 71%|████████████████████████████------------| 5209/7340 [191:02<78:09, 27.3 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 18:37:21,836 - agent.ComputerAgent - INFO - LLM processing started with 6 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 6 messages\n", - "\u001b[92m18:37:21 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 71%|████████████████████████████------------| 5209/7340 [191:03<78:09, 27.3 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m18:37:21 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 18:37:22,505 - agent.ComputerAgent - INFO - Agent: Taking a screenshot to see the current computer screen.\n", - "INFO:agent.ComputerAgent:Agent: Taking a screenshot to see the current computer screen.\n", - "2025-08-11 18:37:22,506 - agent.ComputerAgent - INFO - Computer: click({'x': 1006, 'y': 9})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 1006, 'y': 9})\n", - " 71%|████████████████████████████------------| 5209/7340 [191:04<78:10, 27.3 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 18:37:24,905 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'ctrl+v'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'ctrl+v'})\n", - " 71%|████████████████████████████------------| 5210/7340 [191:06<78:07, 27.3 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/7a6ead00-3730-4f34-9acb-3c8109ec140a/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 18:37:26,064 - agent.ComputerAgent - INFO - LLM processing started with 12 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 12 messages\n", - "\u001b[92m18:37:26 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 18:37:26,733 - agent.ComputerAgent - INFO - LLM processing started with 38 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 38 messages\n", - " 71%|████████████████████████████------------| 5210/7340 [191:08<78:08, 27.3 steps/min]\u001b[92m18:37:26 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m18:37:27 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m18:37:28 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m18:37:28 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/497d5104-1e6e-44a9-a164-fec745a337b6/invoke \"HTTP/1.1 200 OK\"\n", - " 71%|████████████████████████████------------| 5210/7340 [191:10<78:09, 27.3 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m18:37:28 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 18:37:29,395 - agent.ComputerAgent - INFO - Computer: click({'x': 16, 'y': 429})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 16, 'y': 429})\n", - "\u001b[92m18:37:29 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 18:37:30,066 - agent.ComputerAgent - INFO - Computer: click({'x': 18, 'y': 45})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 18, 'y': 45})\n", - "2025-08-11 18:37:30,744 - agent.ComputerAgent - INFO - LLM processing started with 6 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 6 messages\n", - "\u001b[92m18:37:30 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 71%|████████████████████████████------------| 5210/7340 [191:12<78:10, 27.2 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m18:37:30 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 18:37:31,387 - agent.ComputerAgent - INFO - Computer: click({'x': 18, 'y': 239})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 18, 'y': 239})\n", - " 71%|████████████████████████████------------| 5212/7340 [191:13<78:04, 27.3 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m18:37:32 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/create_environment \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - " 71%|████████████████████████████------------| 5213/7340 [191:14<78:01, 27.3 steps/min]\u001b[92m18:37:33 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 18:37:33,218 - agent.ComputerAgent - INFO - Computer: click({'x': 49, 'y': 53})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 49, 'y': 53})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 18:37:34,578 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'enter'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'enter'})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/c1b31663-de2f-4fd6-a091-28bf62a74f86/invoke \"HTTP/1.1 200 OK\"\n", - " 71%|████████████████████████████------------| 5214/7340 [191:16<77:59, 27.3 steps/min]INFO:openai._base_client:Retrying request to /chat/completions in 0.421017 seconds\n", - " 71%|████████████████████████████------------| 5215/7340 [191:17<77:56, 27.3 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/1fef1c7a-93ef-4a63-b067-399dfc4ff08a/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 18:37:36,730 - agent.ComputerAgent - INFO - LLM processing started with 8 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 8 messages\n", - "\u001b[92m18:37:36 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/b3df65c5-9d1c-44fd-b9bb-37f1f0cd64dc/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 71%|████████████████████████████------------| 5215/7340 [191:18<77:57, 27.3 steps/min]2025-08-11 18:37:37,733 - agent.ComputerAgent - INFO - LLM processing started with 8 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 8 messages\n", - "\u001b[92m18:37:37 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/e42da596-e101-4fd3-9dea-8a1d63615dad/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 71%|████████████████████████████------------| 5215/7340 [191:19<77:57, 27.3 steps/min]2025-08-11 18:37:38,404 - agent.ComputerAgent - INFO - LLM processing started with 8 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 8 messages\n", - "\u001b[92m18:37:38 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m18:37:39 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m18:37:39 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/655a0f34-fb5e-49f8-9a65-531af668d6c6/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 71%|████████████████████████████------------| 5215/7340 [191:21<77:58, 27.3 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/7a6ead00-3730-4f34-9acb-3c8109ec140a/invoke \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 18:37:40,395 - agent.ComputerAgent - INFO - LLM processing started with 10 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 10 messages\n", - "\u001b[92m18:37:40 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m18:37:40 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 18:37:41,082 - agent.ComputerAgent - INFO - Computer: click({'x': 605, 'y': 527})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 605, 'y': 527})\n", - "2025-08-11 18:37:41,718 - agent.ComputerAgent - INFO - LLM processing started with 14 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 14 messages\n", - "\u001b[92m18:37:41 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 71%|████████████████████████████------------| 5215/7340 [191:23<77:59, 27.2 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m18:37:41 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 18:37:42,379 - agent.ComputerAgent - INFO - Computer: click({'x': 525, 'y': 502})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 525, 'y': 502})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m18:37:43 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - " 71%|████████████████████████████------------| 5216/7340 [191:25<77:56, 27.2 steps/min]\u001b[92m18:37:43 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m18:37:44 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 18:37:44,917 - agent.ComputerAgent - INFO - Computer: click({'x': 652, 'y': 139})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 652, 'y': 139})\n", - " 71%|████████████████████████████------------| 5217/7340 [191:26<77:54, 27.3 steps/min]\u001b[92m18:37:44 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 18:37:45,530 - agent.ComputerAgent - INFO - Computer: scroll({'scroll_y': -1169, 'scroll_x': 0, 'x': 526, 'y': 427})\n", - "INFO:agent.ComputerAgent:Computer: scroll({'scroll_y': -1169, 'scroll_x': 0, 'x': 526, 'y': 427})\n", - " 71%|████████████████████████████------------| 5219/7340 [191:29<77:49, 27.3 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/3077c8ef-543a-4fa8-b46c-49b632230eed/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m18:37:48 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - " 71%|████████████████████████████------------| 5219/7340 [191:30<77:49, 27.3 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/9b006d7b-b853-41ed-8a84-b7eaa5b6e94b/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 18:37:50,126 - agent.ComputerAgent - INFO - Computer: type({'text': 'Thunderbird'})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': 'Thunderbird'})\n", - "\u001b[92m18:37:50 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v1/environments/1064657b-b89a-4eeb-8197-1c110af6b752/reset \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 71%|████████████████████████████------------| 5219/7340 [191:31<77:50, 27.2 steps/min]2025-08-11 18:37:50,787 - agent.ComputerAgent - INFO - Computer: click({'x': 749, 'y': 440})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 749, 'y': 440})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/055914cd-07b0-4dcd-9407-c6975b1eccbf/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 18:37:51,454 - agent.ComputerAgent - INFO - LLM processing started with 40 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 40 messages\n", - "\u001b[92m18:37:51 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/3077c8ef-543a-4fa8-b46c-49b632230eed/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m18:37:52 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 71%|████████████████████████████------------| 5220/7340 [191:34<77:48, 27.2 steps/min]\u001b[92m18:37:52 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 18:37:53,444 - agent.ComputerAgent - INFO - LLM processing started with 42 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 42 messages\n", - "\u001b[92m18:37:53 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m18:37:53 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 18:37:54,116 - agent.ComputerAgent - INFO - Computer: click({'x': 1008, 'y': 10})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 1008, 'y': 10})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/a72854f0-3bb0-4711-a18e-7a467a56390e/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/3077c8ef-543a-4fa8-b46c-49b632230eed/close \"HTTP/1.1 200 OK\"\n", - " 71%|████████████████████████████------------| 5223/7340 [191:35<77:39, 27.3 steps/min]\u001b[92m18:37:54 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 18:37:55,819 - agent.ComputerAgent - INFO - Computer: click({'x': 80, 'y': 430})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 80, 'y': 430})\n", - "2025-08-11 18:37:56,505 - agent.ComputerAgent - INFO - LLM processing started with 8 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 8 messages\n", - "\u001b[92m18:37:56 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/1fef1c7a-93ef-4a63-b067-399dfc4ff08a/invoke \"HTTP/1.1 200 OK\"\n", - " 71%|████████████████████████████------------| 5224/7340 [191:38<77:37, 27.3 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m18:37:57 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 18:37:57,824 - agent.ComputerAgent - INFO - LLM processing started with 10 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 10 messages\n", - "\u001b[92m18:37:57 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 71%|████████████████████████████------------| 5225/7340 [191:39<77:34, 27.3 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/e42da596-e101-4fd3-9dea-8a1d63615dad/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: GET https://orchestration.hud.so/hud-gym/api/v1/gyms/OSWorld-Ubuntu \"HTTP/1.1 200 OK\"\n", - "Loading checkpoint shards: 0%| | 0/4 [00:00\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/af5d4b08-d761-4bf8-a2c7-7ce16ed98ab9/invoke \"HTTP/1.1 200 OK\"\n", - "Loading checkpoint shards: 0%| | 0/4 [00:00 Settings > On startup.\n", - "- Select “Open the New Tab page,” or “Continue where you left off.”\n", - "- If “Open a specific page or set of pages” is selected, remove funbrain.com (click the three dots beside it > Remove).\n", - "- Also check Settings > Appearance > Show home button. If enabled, set it to “New Tab page” and make sure funbrain.com isn’t listed there.\n", - "\n", - "2) Unpin any pinned tab\n", - "- If you see a tiny pin icon on the funbrain tab, right‑click the tab > Unpin. Close it.\n", - "\n", - "3) Remove any extension hijacking your homepage/new tab\n", - "- Go to chrome://extensions\n", - "- Toggle off or Remove any unknown extensions, especially ones referencing “New Tab,” “Home,” “Search,” or anything you don’t recognize.\n", - "- Restart Chrome.\n", - "\n", - "4) Check if Chrome is “Managed by your organization” (policy hijack)\n", - "- Go to chrome://policy\n", - "- If you see policies like HomepageLocation, NewTabPageLocation, RestoreOnStartup, or StartupURLs pointing to funbrain.com (and this is your personal device), remove the policies:\n", - " - Press Windows key, type “regedit”, open Registry Editor.\n", - " - Navigate to:\n", - " - HKEY_CURRENT_USER\\Software\\Policies\\Google\\Chrome\n", - " - HKEY_LOCAL_MACHINE\\SOFTWARE\\Policies\\Google\\Chrome\n", - " - Delete values like HomepageLocation, StartupURLs, RestoreOnStartup, NewTabPageLocation, or delete the Chrome key entirely if it only contains unwanted values.\n", - " - Easier method (PowerShell as Administrator):\n", - " Remove-Item -Path \"HKCU:\\Software\\Policies\\Google\\Chrome\" -Recurse -Force -ErrorAction SilentlyContinue\n", - " Remove-Item -Path \"HKLM:\\SOFTWARE\\Policies\\Google\\Chrome\" -Recurse -Force -ErrorAction SilentlyContinue\n", - " - Reboot and recheck chrome://policy (it should be empty or unrelated).\n", - "\n", - "5) Fix the shortcut that launches Chrome\n", - "- Right‑click the Chrome shortcut you normall...\n", - "INFO:agent.ComputerAgent:Agent: This isn’t a cache issue. Chrome is being told to open that site by a setting, an extension/policy, or a modified shortcut. Do the steps below in order (quickest first). They’re written for Windows on your Surface Pro.\n", - "\n", - "1) Fix Chrome’s startup settings\n", - "- Open Chrome.\n", - "- Menu (three dots) > Settings > On startup.\n", - "- Select “Open the New Tab page,” or “Continue where you left off.”\n", - "- If “Open a specific page or set of pages” is selected, remove funbrain.com (click the three dots beside it > Remove).\n", - "- Also check Settings > Appearance > Show home button. If enabled, set it to “New Tab page” and make sure funbrain.com isn’t listed there.\n", - "\n", - "2) Unpin any pinned tab\n", - "- If you see a tiny pin icon on the funbrain tab, right‑click the tab > Unpin. Close it.\n", - "\n", - "3) Remove any extension hijacking your homepage/new tab\n", - "- Go to chrome://extensions\n", - "- Toggle off or Remove any unknown extensions, especially ones referencing “New Tab,” “Home,” “Search,” or anything you don’t recognize.\n", - "- Restart Chrome.\n", - "\n", - "4) Check if Chrome is “Managed by your organization” (policy hijack)\n", - "- Go to chrome://policy\n", - "- If you see policies like HomepageLocation, NewTabPageLocation, RestoreOnStartup, or StartupURLs pointing to funbrain.com (and this is your personal device), remove the policies:\n", - " - Press Windows key, type “regedit”, open Registry Editor.\n", - " - Navigate to:\n", - " - HKEY_CURRENT_USER\\Software\\Policies\\Google\\Chrome\n", - " - HKEY_LOCAL_MACHINE\\SOFTWARE\\Policies\\Google\\Chrome\n", - " - Delete values like HomepageLocation, StartupURLs, RestoreOnStartup, NewTabPageLocation, or delete the Chrome key entirely if it only contains unwanted values.\n", - " - Easier method (PowerShell as Administrator):\n", - " Remove-Item -Path \"HKCU:\\Software\\Policies\\Google\\Chrome\" -Recurse -Force -ErrorAction SilentlyContinue\n", - " Remove-Item -Path \"HKLM:\\SOFTWARE\\Policies\\Google\\Chrome\" -Recurse -Force -ErrorAction SilentlyContinue\n", - " - Reboot and recheck chrome://policy (it should be empty or unrelated).\n", - "\n", - "5) Fix the shortcut that launches Chrome\n", - "- Right‑click the Chrome shortcut you normall...\n", - "2025-08-11 18:49:11,237 - agent.ComputerAgent - INFO - Total usage:\n", - " - completion_tokens: 2424\n", - " - prompt_tokens: 1953\n", - " - total_tokens: 4377\n", - " - completion_tokens_details:\n", - " - accepted_prediction_tokens: 0\n", - " - audio_tokens: 0\n", - " - reasoning_tokens: 1600\n", - " - rejected_prediction_tokens: 0\n", - " - prompt_tokens_details:\n", - " - audio_tokens: 0\n", - " - cached_tokens: 0\n", - " - response_cost: $0.0267\n", - "INFO:agent.ComputerAgent:Total usage:\n", - " - completion_tokens: 2424\n", - " - prompt_tokens: 1953\n", - " - total_tokens: 4377\n", - " - completion_tokens_details:\n", - " - accepted_prediction_tokens: 0\n", - " - audio_tokens: 0\n", - " - reasoning_tokens: 1600\n", - " - rejected_prediction_tokens: 0\n", - " - prompt_tokens_details:\n", - " - audio_tokens: 0\n", - " - cached_tokens: 0\n", - " - response_cost: $0.0267\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/497d5104-1e6e-44a9-a164-fec745a337b6/invoke \"HTTP/1.1 200 OK\"\n", - " 76%|██████████████████████████████----------| 5568/7340 [202:52<64:34, 27.4 steps/min]2025-08-11 18:49:11,903 - agent.ComputerAgent - INFO - LLM processing started with 30 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 30 messages\n", - "\u001b[92m18:49:11 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 18:49:12,597 - agent.ComputerAgent - INFO - LLM processing started with 36 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 36 messages\n", - "\u001b[92m18:49:12 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/0ce9d177-2b9a-4fde-a8a5-eb1b59248c8f/invoke \"HTTP/1.1 200 OK\"\n", - " 76%|██████████████████████████████----------| 5568/7340 [202:54<64:34, 27.4 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m18:49:13 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 18:49:13,963 - agent.ComputerAgent - INFO - LLM processing started with 38 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 38 messages\n", - "\u001b[92m18:49:14 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 76%|██████████████████████████████----------| 5568/7340 [202:55<64:34, 27.4 steps/min]\u001b[92m18:49:14 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 18:49:14,626 - agent.ComputerAgent - INFO - Computer: click({'button': 'left', 'x': 18, 'y': 237})\n", - "INFO:agent.ComputerAgent:Computer: click({'button': 'left', 'x': 18, 'y': 237})\n", - "2025-08-11 18:49:16,032 - agent.ComputerAgent - INFO - LLM processing started with 30 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 30 messages\n", - "\u001b[92m18:49:16 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/8a9dd85a-f951-495e-aea0-d3864853591e/invoke \"HTTP/1.1 200 OK\"\n", - " 76%|██████████████████████████████----------| 5568/7340 [202:57<64:35, 27.4 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 18:49:17,933 - agent.ComputerAgent - INFO - Computer: type({'text': 'ICLR 2013 location'})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': 'ICLR 2013 location'})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 18:49:19,279 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'ctrl+a'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'ctrl+a'})\n", - " 76%|██████████████████████████████----------| 5569/7340 [203:01<64:33, 27.4 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/8a9dd85a-f951-495e-aea0-d3864853591e/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m18:49:19 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 18:49:21,241 - agent.ComputerAgent - INFO - Agent: Taking a screenshot to see the current computer screen.\n", - "INFO:agent.ComputerAgent:Agent: Taking a screenshot to see the current computer screen.\n", - "2025-08-11 18:49:21,241 - agent.ComputerAgent - INFO - Agent: Yes. VLC is set by default to reuse the first instance when you open a file (especially from the file manager). Turn that off and you can run multiple, independent players.\n", - "\n", - "Do this:\n", - "- Open VLC.\n", - "- Go to Tools -> Preferences (Ctrl+P).\n", - "- In the Simple settings view on the Interface tab, find Playlist and Instances.\n", - "- Uncheck:\n", - " - Allow only one instance\n", - " - Use only one instance when started from file manager\n", - " - Enqueue items into playlist in one instance (if you see it)\n", - "- Click Save.\n", - "- Quit all VLC windows completely (Ctrl+Q) so the setting takes effect, then start VLC again and open a different file in each window.\n", - "\n", - "Command-line alternative (Windows/macOS/Linux):\n", - "- Start each player with:\n", - " - vlc --no-one-instance --no-one-instance-when-started-from-file-manager file1.mp4\n", - " - vlc --no-one-instance file2.mp4\n", - "\n", - "After this, each VLC window will play its own video simultaneously.\n", - "\n", - "Task completed\n", - "INFO:agent.ComputerAgent:Agent: Yes. VLC is set by default to reuse the first instance when you open a file (especially from the file manager). Turn that off and you can run multiple, independent players.\n", - "\n", - "Do this:\n", - "- Open VLC.\n", - "- Go to Tools -> Preferences (Ctrl+P).\n", - "- In the Simple settings view on the Interface tab, find Playlist and Instances.\n", - "- Uncheck:\n", - " - Allow only one instance\n", - " - Use only one instance when started from file manager\n", - " - Enqueue items into playlist in one instance (if you see it)\n", - "- Click Save.\n", - "- Quit all VLC windows completely (Ctrl+Q) so the setting takes effect, then start VLC again and open a different file in each window.\n", - "\n", - "Command-line alternative (Windows/macOS/Linux):\n", - "- Start each player with:\n", - " - vlc --no-one-instance --no-one-instance-when-started-from-file-manager file1.mp4\n", - " - vlc --no-one-instance file2.mp4\n", - "\n", - "After this, each VLC window will play its own video simultaneously.\n", - "\n", - "Task completed\n", - "2025-08-11 18:49:21,909 - agent.ComputerAgent - INFO - Total usage:\n", - " - completion_tokens: 1307\n", - " - prompt_tokens: 1975\n", - " - total_tokens: 3282\n", - " - completion_tokens_details:\n", - " - accepted_prediction_tokens: 0\n", - " - audio_tokens: 0\n", - " - reasoning_tokens: 1088\n", - " - rejected_prediction_tokens: 0\n", - " - prompt_tokens_details:\n", - " - audio_tokens: 0\n", - " - cached_tokens: 0\n", - " - response_cost: $0.0155\n", - "INFO:agent.ComputerAgent:Total usage:\n", - " - completion_tokens: 1307\n", - " - prompt_tokens: 1975\n", - " - total_tokens: 3282\n", - " - completion_tokens_details:\n", - " - accepted_prediction_tokens: 0\n", - " - audio_tokens: 0\n", - " - reasoning_tokens: 1088\n", - " - rejected_prediction_tokens: 0\n", - " - prompt_tokens_details:\n", - " - audio_tokens: 0\n", - " - cached_tokens: 0\n", - " - response_cost: $0.0155\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/655a0f34-fb5e-49f8-9a65-531af668d6c6/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m18:49:22 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 76%|██████████████████████████████----------| 5590/7340 [203:04<63:34, 27.5 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/35ee5a0d-d8f9-4419-b253-d394f35fe993/invoke \"HTTP/1.1 200 OK\"\n", - "\u001b[92m18:49:22 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m18:49:22 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v1/environments/0835b0ab-8369-4209-a85b-955dbb271b54/reset \"HTTP/1.1 200 OK\"\n", - "2025-08-11 18:49:23,196 - agent.ComputerAgent - INFO - LLM processing started with 8 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 8 messages\n", - "\u001b[92m18:49:23 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m18:49:23 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 18:49:23,915 - agent.ComputerAgent - INFO - Computer: click({'x': 109, 'y': 125})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 109, 'y': 125})\n", - "\u001b[92m18:49:23 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/8a9dd85a-f951-495e-aea0-d3864853591e/close \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/655a0f34-fb5e-49f8-9a65-531af668d6c6/close \"HTTP/1.1 200 OK\"\n", - " 76%|██████████████████████████████----------| 5591/7340 [203:05<63:31, 27.5 steps/min]2025-08-11 18:49:24,590 - agent.ComputerAgent - INFO - Computer: drag({'path': [{'x': 75, 'y': 177}, {'x': 278, 'y': 177}]})\n", - "INFO:agent.ComputerAgent:Computer: drag({'path': [{'x': 75, 'y': 177}, {'x': 278, 'y': 177}]})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/aa800986-7030-4845-b4a1-82119abb97e9/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/84265bb9-b6f6-479e-8a58-920cfa2b7c69/invoke \"HTTP/1.1 200 OK\"\n", - " 76%|██████████████████████████████----------| 5592/7340 [203:07<63:29, 27.5 steps/min]INFO:httpx:HTTP Request: GET https://orchestration.hud.so/hud-gym/api/v1/gyms/OSWorld-Ubuntu \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: GET https://orchestration.hud.so/hud-gym/api/v1/gyms/OSWorld-Ubuntu \"HTTP/1.1 200 OK\"\n", - " 76%|██████████████████████████████----------| 5600/7340 [203:08<63:07, 27.6 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 18:49:28,277 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'ctrl+c'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'ctrl+c'})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/0835b0ab-8369-4209-a85b-955dbb271b54/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/a72854f0-3bb0-4711-a18e-7a467a56390e/invoke \"HTTP/1.1 200 OK\"\n", - " 76%|██████████████████████████████----------| 5600/7340 [203:10<63:07, 27.6 steps/min]2025-08-11 18:49:28,957 - agent.ComputerAgent - INFO - LLM processing started with 1 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 1 messages\n", - "\u001b[92m18:49:28 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m18:49:29 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 18:49:30,327 - agent.ComputerAgent - INFO - LLM processing started with 32 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 32 messages\n", - "\u001b[92m18:49:30 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m18:49:31 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/84265bb9-b6f6-479e-8a58-920cfa2b7c69/close \"HTTP/1.1 200 OK\"\n", - " 76%|██████████████████████████████----------| 5600/7340 [203:12<63:08, 27.6 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/1064657b-b89a-4eeb-8197-1c110af6b752/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/584f1ba5-3dc8-4b11-9242-7100c4e1133e/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/aa800986-7030-4845-b4a1-82119abb97e9/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 18:49:32,369 - agent.ComputerAgent - INFO - LLM processing started with 40 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 40 messages\n", - "\u001b[92m18:49:32 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "Loading checkpoint shards: 0%| | 0/4 [00:00 log.txt\\n'})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': 'cd ~/Desktop\\npython3 calculator.py > log.txt\\n'})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/10c454d7-987f-4a23-83d6-534bd9ba42c2/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - " 83%|█████████████████████████████████-------| 6089/7340 [220:38<45:19, 27.6 steps/min]\u001b[92m19:06:56 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/b641dbb7-3e3c-437d-bc11-5e038171855d/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/43a383a0-163d-4a8b-8494-0e1d1eab6cd6/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/27fc4825-1617-494a-9308-b128bd8af05a/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:06:57,016 - agent.ComputerAgent - INFO - LLM processing started with 6 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 6 messages\n", - "\u001b[92m19:06:57 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m19:06:57 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/80d19a15-b1ca-43cc-8d1b-1f86242172b5/invoke \"HTTP/1.1 200 OK\"\n", - "Loading checkpoint shards: 0%| | 0/4 [00:00 Modify > Font > choose Liberation Sans Narrow > OK.\n", - "- Drawing styles (for manual text boxes):\n", - " - Click the Drawing Styles icon.\n", - " - Right‑click Default > Modify > Font > Liberation Sans Narrow > OK.\n", - "- Remove any direct overrides so objects follow the styles:\n", - " - In the Slides pane, select all slides (click first, Shift‑click last).\n", - " - Press Ctrl+M (Format > Clear Direct Formatting). This makes existing text boxes pick up the new style font.\n", - "\n", - "Method 2: One‑click change with a macro (works even if text has direct formatting)\n", - "- Tools > Macros > Organize Macros > LibreOffice Basic > select your presentation > New.\n", - "- Paste this macro into the new module and Run:\n", - "\n", - "Sub SetAllTextToLiberationNarrow\n", - " Dim oDoc As Object, oSlides As Object, oSlide As Object\n", - " Dim oShape As Object\n", - " oDoc = ThisComponent\n", - " oSlides = oDoc.getDrawPages()\n", - " Dim i As Long, j As Long\n", - " For i = 0 To oSlides.getCount() - 1\n", - " oSlide = oSlides.getByIndex(i)\n", - " For j = 0 To oSlide.getCount() - 1\n", - " oShape = oSlide.getByIndex(j)\n", - " Call SetFontRecursive(oShape)\n", - " Next j\n", - " Next i\n", - "End Sub\n", - "\n", - "Sub SetFontRecursive(oShape As Object)\n", - " If oShape.supportsService(\"com.sun.star.drawing.GroupShape\") Then\n", - " Dim k As Long\n", - " For k = 0 To oShape.getCount() - 1\n", - " Call SetFontRecursive(oShape.getByIndex(k))\n", - " Next k\n", - " ElseIf oShape.supportsService(\"com.sun.star.drawing.Text\") Then\n", - " Dim oText As Object, oCur As Object\n", - " oText = oShape.getText()\n", - " oCur = oText.createTextCursor()\n", - " oCur.CharFontName = \"Liberation Sans Narrow\"\n", - " oCur.CharF...\n", - "INFO:agent.ComputerAgent:Agent: Yes—here are reliable ways to standardize every text box to Liberation Sans Narrow in LibreOffice Impress without clicking each one.\n", - "\n", - "Method 1 (recommended): Change styles so all objects inherit the font\n", - "- Open your presentation.\n", - "- Press F11 to open the Styles deck.\n", - "- Presentation styles (for placeholders):\n", - " - Click the Presentation Styles icon.\n", - " - Right‑click each of Title, Subtitle, Outline 1–10 (or Outline), Notes, etc. > Modify > Font > choose Liberation Sans Narrow > OK.\n", - "- Drawing styles (for manual text boxes):\n", - " - Click the Drawing Styles icon.\n", - " - Right‑click Default > Modify > Font > Liberation Sans Narrow > OK.\n", - "- Remove any direct overrides so objects follow the styles:\n", - " - In the Slides pane, select all slides (click first, Shift‑click last).\n", - " - Press Ctrl+M (Format > Clear Direct Formatting). This makes existing text boxes pick up the new style font.\n", - "\n", - "Method 2: One‑click change with a macro (works even if text has direct formatting)\n", - "- Tools > Macros > Organize Macros > LibreOffice Basic > select your presentation > New.\n", - "- Paste this macro into the new module and Run:\n", - "\n", - "Sub SetAllTextToLiberationNarrow\n", - " Dim oDoc As Object, oSlides As Object, oSlide As Object\n", - " Dim oShape As Object\n", - " oDoc = ThisComponent\n", - " oSlides = oDoc.getDrawPages()\n", - " Dim i As Long, j As Long\n", - " For i = 0 To oSlides.getCount() - 1\n", - " oSlide = oSlides.getByIndex(i)\n", - " For j = 0 To oSlide.getCount() - 1\n", - " oShape = oSlide.getByIndex(j)\n", - " Call SetFontRecursive(oShape)\n", - " Next j\n", - " Next i\n", - "End Sub\n", - "\n", - "Sub SetFontRecursive(oShape As Object)\n", - " If oShape.supportsService(\"com.sun.star.drawing.GroupShape\") Then\n", - " Dim k As Long\n", - " For k = 0 To oShape.getCount() - 1\n", - " Call SetFontRecursive(oShape.getByIndex(k))\n", - " Next k\n", - " ElseIf oShape.supportsService(\"com.sun.star.drawing.Text\") Then\n", - " Dim oText As Object, oCur As Object\n", - " oText = oShape.getText()\n", - " oCur = oText.createTextCursor()\n", - " oCur.CharFontName = \"Liberation Sans Narrow\"\n", - " oCur.CharF...\n", - "2025-08-11 19:26:25,833 - agent.ComputerAgent - INFO - Total usage:\n", - " - completion_tokens: 3294\n", - " - prompt_tokens: 1964\n", - " - total_tokens: 5258\n", - " - completion_tokens_details:\n", - " - accepted_prediction_tokens: 0\n", - " - audio_tokens: 0\n", - " - reasoning_tokens: 2560\n", - " - rejected_prediction_tokens: 0\n", - " - prompt_tokens_details:\n", - " - audio_tokens: 0\n", - " - cached_tokens: 0\n", - " - response_cost: $0.0354\n", - "INFO:agent.ComputerAgent:Total usage:\n", - " - completion_tokens: 3294\n", - " - prompt_tokens: 1964\n", - " - total_tokens: 5258\n", - " - completion_tokens_details:\n", - " - accepted_prediction_tokens: 0\n", - " - audio_tokens: 0\n", - " - reasoning_tokens: 2560\n", - " - rejected_prediction_tokens: 0\n", - " - prompt_tokens_details:\n", - " - audio_tokens: 0\n", - " - cached_tokens: 0\n", - " - response_cost: $0.0354\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/7a2c2835-d21e-4e04-babb-e8305a4f1f9d/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/1a178f89-87e5-46d9-a114-22d5fcc5c630/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/6f055f32-c606-4a58-91bc-c5990d4e8280/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/f209659e-2ed9-40fa-ae61-5359fb9ef290/invoke \"HTTP/1.1 200 OK\"\n", - " 90%|████████████████████████████████████----| 6611/7340 [240:07<26:28, 27.5 steps/min]2025-08-11 19:26:26,517 - agent.ComputerAgent - INFO - Computer: double_click({'x': 884, 'y': 123})\n", - "INFO:agent.ComputerAgent:Computer: double_click({'x': 884, 'y': 123})\n", - "2025-08-11 19:26:27,223 - agent.ComputerAgent - INFO - Computer: click({'x': 344, 'y': 34})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 344, 'y': 34})\n", - "2025-08-11 19:26:27,897 - agent.ComputerAgent - INFO - Computer: click({'x': 548, 'y': 249})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 548, 'y': 249})\n", - "2025-08-11 19:26:28,563 - agent.ComputerAgent - INFO - LLM processing started with 10 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 10 messages\n", - "\u001b[92m19:26:28 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 90%|████████████████████████████████████----| 6613/7340 [240:10<26:24, 27.5 steps/min]2025-08-11 19:26:29,242 - agent.ComputerAgent - INFO - LLM processing started with 26 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 26 messages\n", - "\u001b[92m19:26:29 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:26:30,582 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'f1'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'f1'})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:26:31,909 - agent.ComputerAgent - INFO - Computer: wait({})\n", - "INFO:agent.ComputerAgent:Computer: wait({})\n", - " 90%|████████████████████████████████████----| 6616/7340 [240:13<26:17, 27.5 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:26:33,178 - agent.ComputerAgent - INFO - Computer: screenshot({})\n", - "INFO:agent.ComputerAgent:Computer: screenshot({})\n", - " 90%|████████████████████████████████████----| 6618/7340 [240:14<26:12, 27.5 steps/min]2025-08-11 19:26:33,810 - agent.ComputerAgent - INFO - LLM processing started with 1 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 1 messages\n", - "\u001b[92m19:26:33 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 19:26:34,481 - agent.ComputerAgent - INFO - LLM processing started with 8 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 8 messages\n", - "\u001b[92m19:26:34 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/09c933ad-61bf-4498-b248-0df86e3aea78/invoke \"HTTP/1.1 200 OK\"\n", - " 90%|████████████████████████████████████----| 6619/7340 [240:16<26:10, 27.5 steps/min]2025-08-11 19:26:35,113 - agent.ComputerAgent - INFO - LLM processing started with 8 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 8 messages\n", - "\u001b[92m19:26:35 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 19:26:35,982 - agent.ComputerAgent - INFO - LLM processing started with 14 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 14 messages\n", - "\u001b[92m19:26:36 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/7f4008ee-6c98-4905-9ade-965ea7842b64/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 90%|████████████████████████████████████----| 6619/7340 [240:17<26:10, 27.5 steps/min]2025-08-11 19:26:37,325 - agent.ComputerAgent - INFO - LLM processing started with 38 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 38 messages\n", - "\u001b[92m19:26:37 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/18debd9e-6c58-4504-8a04-13cba683a254/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:26:38 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 90%|████████████████████████████████████----| 6619/7340 [240:19<26:10, 27.5 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/2d8a6e51-acdb-47b9-8ee4-f3085c741fd5/invoke \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:26:38,682 - agent.ComputerAgent - INFO - LLM processing started with 40 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 40 messages\n", - "\u001b[92m19:26:38 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/bcec4523-df7a-48b5-aea1-8d7c632a6dc4/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/51954fb4-34ed-4511-b2fd-a6169b5ea5d3/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/9a544504-3e48-48b2-8429-0a97e266ebfb/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/84a5d283-63f1-43fc-b483-76116d67f385/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/c915dbd9-32bc-40a7-9c07-d437c737419f/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:26:39,380 - agent.ComputerAgent - INFO - LLM processing started with 36 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 36 messages\n", - "\u001b[92m19:26:39 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m19:26:39 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/775a5b67-2406-42b8-86e5-243e01b8dc27/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/052ac585-1998-46b2-9ac5-0dc192aeba02/invoke \"HTTP/1.1 200 OK\"\n", - " 90%|████████████████████████████████████----| 6619/7340 [240:21<26:10, 27.5 steps/min]" - ] - }, - { - "name": "stdout", - "output_type": "stream", - "text": [ - "No screenshot found, taking screenshot\n" - ] - }, - { - "name": "stderr", - "output_type": "stream", - "text": [ - "2025-08-11 19:26:40,063 - agent.ComputerAgent - INFO - LLM processing started with 8 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 8 messages\n", - "\u001b[92m19:26:40 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 19:26:40,762 - agent.ComputerAgent - INFO - Computer: click({'button': 'left', 'x': 344, 'y': 137})\n", - "INFO:agent.ComputerAgent:Computer: click({'button': 'left', 'x': 344, 'y': 137})\n", - " 90%|████████████████████████████████████----| 6619/7340 [240:22<26:11, 27.5 steps/min]2025-08-11 19:26:41,401 - agent.ComputerAgent - INFO - LLM processing started with 32 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 32 messages\n", - "\u001b[92m19:26:41 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 19:26:42,073 - agent.ComputerAgent - INFO - LLM processing started with 38 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 38 messages\n", - "\u001b[92m19:26:42 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 19:26:42,723 - agent.ComputerAgent - INFO - LLM processing started with 30 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 30 messages\n", - "\u001b[92m19:26:42 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 90%|████████████████████████████████████----| 6620/7340 [240:24<26:08, 27.5 steps/min]2025-08-11 19:26:43,402 - agent.ComputerAgent - INFO - LLM processing started with 26 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 26 messages\n", - "\u001b[92m19:26:43 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:26:44 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:26:45,407 - agent.ComputerAgent - INFO - Computer: wait({})\n", - "INFO:agent.ComputerAgent:Computer: wait({})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:26:46 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:26:46 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 90%|████████████████████████████████████----| 6620/7340 [240:29<26:09, 27.5 steps/min]\u001b[92m19:26:47 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m19:26:47 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:26:48,083 - agent.ComputerAgent - INFO - LLM processing started with 38 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 38 messages\n", - "\u001b[92m19:26:48 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:26:48,753 - agent.ComputerAgent - INFO - Computer: click({'x': 914, 'y': 660})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 914, 'y': 660})\n", - "\u001b[92m19:26:48 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - " 90%|████████████████████████████████████----| 6621/7340 [240:31<26:07, 27.5 steps/min]\u001b[92m19:26:49 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:26:50,093 - agent.ComputerAgent - INFO - Computer: click({'x': 641, 'y': 498})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 641, 'y': 498})\n", - "\u001b[92m19:26:50 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m19:26:50 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:26:50,755 - agent.ComputerAgent - INFO - Computer: click({'x': 893, 'y': 134})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 893, 'y': 134})\n", - "2025-08-11 19:26:51,455 - agent.ComputerAgent - INFO - Agent: Taking a screenshot to see the current computer screen.\n", - "INFO:agent.ComputerAgent:Agent: Taking a screenshot to see the current computer screen.\n", - "2025-08-11 19:26:51,456 - agent.ComputerAgent - INFO - Computer: click({'button': 'left', 'x': 386, 'y': 250})\n", - "INFO:agent.ComputerAgent:Computer: click({'button': 'left', 'x': 386, 'y': 250})\n", - "\u001b[92m19:26:51 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 90%|████████████████████████████████████----| 6622/7340 [240:33<26:04, 27.5 steps/min]2025-08-11 19:26:52,153 - agent.ComputerAgent - INFO - Computer: click({'x': 173, 'y': 150})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 173, 'y': 150})\n", - "2025-08-11 19:26:52,821 - agent.ComputerAgent - INFO - LLM processing started with 16 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 16 messages\n", - "\u001b[92m19:26:52 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 90%|████████████████████████████████████----| 6626/7340 [240:35<25:55, 27.5 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:26:55 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 90%|████████████████████████████████████----| 6626/7340 [240:36<25:55, 27.5 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/77892268-14f2-4dfa-b58c-6a682f258679/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:26:56 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m19:26:56 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 90%|████████████████████████████████████----| 6626/7340 [240:37<25:55, 27.5 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:26:56,855 - agent.ComputerAgent - INFO - Computer: click({'x': 46, 'y': 528})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 46, 'y': 528})\n", - "\u001b[92m19:26:56 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:26:57,527 - agent.ComputerAgent - INFO - Computer: click({'x': 731, 'y': 617})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 731, 'y': 617})\n", - "2025-08-11 19:26:58,183 - agent.ComputerAgent - INFO - LLM processing started with 34 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 34 messages\n", - "\u001b[92m19:26:58 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 90%|████████████████████████████████████----| 6626/7340 [240:39<25:56, 27.5 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/7a2c2835-d21e-4e04-babb-e8305a4f1f9d/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/6f055f32-c606-4a58-91bc-c5990d4e8280/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/1a178f89-87e5-46d9-a114-22d5fcc5c630/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:26:58,873 - agent.ComputerAgent - INFO - LLM processing started with 12 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 12 messages\n", - "\u001b[92m19:26:58 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/73c70c0d-c1a0-401f-83c0-063e983abd6c/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/af58ffed-65a3-4c4a-a9fe-5c940230627d/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/f209659e-2ed9-40fa-ae61-5359fb9ef290/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:26:59,543 - agent.ComputerAgent - INFO - LLM processing started with 28 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 28 messages\n", - "\u001b[92m19:26:59 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:27:00,884 - agent.ComputerAgent - INFO - Computer: type({'text': '20'})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': '20'})\n", - " 90%|████████████████████████████████████----| 6628/7340 [240:42<25:51, 27.5 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/7f4008ee-6c98-4905-9ade-965ea7842b64/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:27:01,553 - agent.ComputerAgent - INFO - LLM processing started with 16 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 16 messages\n", - "\u001b[92m19:27:01 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 19:27:02,260 - agent.ComputerAgent - INFO - LLM processing started with 6 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 6 messages\n", - "\u001b[92m19:27:02 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 19:27:02,945 - agent.ComputerAgent - INFO - LLM processing started with 10 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 10 messages\n", - "\u001b[92m19:27:02 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/7f4008ee-6c98-4905-9ade-965ea7842b64/close \"HTTP/1.1 200 OK\"\n", - " 91%|████████████████████████████████████----| 6648/7340 [240:44<25:03, 27.6 steps/min]2025-08-11 19:27:04,256 - agent.ComputerAgent - INFO - LLM processing started with 10 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 10 messages\n", - "\u001b[92m19:27:04 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 91%|████████████████████████████████████----| 6648/7340 [240:46<25:03, 27.6 steps/min]INFO:httpx:HTTP Request: GET https://orchestration.hud.so/hud-gym/api/v1/gyms/OSWorld-Ubuntu \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/2d8a6e51-acdb-47b9-8ee4-f3085c741fd5/invoke \"HTTP/1.1 200 OK\"\n", - " 91%|████████████████████████████████████----| 6648/7340 [240:47<25:03, 27.6 steps/min]2025-08-11 19:27:05,962 - agent.ComputerAgent - INFO - LLM processing started with 42 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 42 messages\n", - "\u001b[92m19:27:05 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 91%|████████████████████████████████████----| 6648/7340 [240:48<25:03, 27.6 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:27:07,803 - agent.ComputerAgent - INFO - Computer: type({'text': 'Extensions: Install from VSIX'})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': 'Extensions: Install from VSIX'})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/775a5b67-2406-42b8-86e5-243e01b8dc27/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:27:08 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:27:09,824 - agent.ComputerAgent - INFO - Computer: get_current_url({})\n", - "INFO:agent.ComputerAgent:Computer: get_current_url({})\n", - " 91%|████████████████████████████████████----| 6648/7340 [240:51<25:04, 27.6 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/18debd9e-6c58-4504-8a04-13cba683a254/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Loading checkpoint shards: 0%| | 0/4 [00:00\n", - "Loading checkpoint shards: 0%| | 0/4 [00:00 View > Folders > Unified.\n", - "- In the left sidebar you’ll now see a “Unified Folders” section with a single “Inbox” that aggregates all accounts.\n", - "\n", - "Optional: choose exactly which accounts/folders are included\n", - "- In the left sidebar, under Unified Folders, right‑click Inbox > Properties.\n", - "- Click “Select the folders to search,” then check the Inbox for each account you want included. Click Update.\n", - "\n", - "Notes and tips\n", - "- Thunderbird remembers this view; if it ever changes, repeat View > Folders > Unified.\n", - "- If you use POP and want all mail to physically go to one inbox, set a Global Inbox: Account Settings > your POP account > Server Settings > Advanced > Global Inbox (Local Folders).\n", - "- If a folder is missing from the Unified view, right‑click that folder > Subscribe (for IMAP) to ensure it’s subscribed, and confirm it’s checked in the Unified Inbox Properties.\n", - "\n", - "Task completed\n", - "INFO:agent.ComputerAgent:Agent: Here’s how to set up a unified inbox in Thunderbird so all your accounts’ emails show in one place:\n", - "\n", - "- Open Thunderbird.\n", - "- Go to the application menu (three horizontal lines in the top-right) > View > Folders > Unified.\n", - "- In the left sidebar you’ll now see a “Unified Folders” section with a single “Inbox” that aggregates all accounts.\n", - "\n", - "Optional: choose exactly which accounts/folders are included\n", - "- In the left sidebar, under Unified Folders, right‑click Inbox > Properties.\n", - "- Click “Select the folders to search,” then check the Inbox for each account you want included. Click Update.\n", - "\n", - "Notes and tips\n", - "- Thunderbird remembers this view; if it ever changes, repeat View > Folders > Unified.\n", - "- If you use POP and want all mail to physically go to one inbox, set a Global Inbox: Account Settings > your POP account > Server Settings > Advanced > Global Inbox (Local Folders).\n", - "- If a folder is missing from the Unified view, right‑click that folder > Subscribe (for IMAP) to ensure it’s subscribed, and confirm it’s checked in the Unified Inbox Properties.\n", - "\n", - "Task completed\n", - "Loading checkpoint shards: 25%|██▌ | 1/4 [00:01<00:05, 1.72s/it]2025-08-11 19:29:25,906 - agent.ComputerAgent - INFO - Total usage:\n", - " - completion_tokens: 1266\n", - " - prompt_tokens: 12578\n", - " - total_tokens: 13844\n", - " - completion_tokens_details:\n", - " - accepted_prediction_tokens: 0\n", - " - audio_tokens: 0\n", - " - reasoning_tokens: 1024\n", - " - rejected_prediction_tokens: 0\n", - " - prompt_tokens_details:\n", - " - audio_tokens: 0\n", - " - cached_tokens: 11776\n", - " - response_cost: $0.0151\n", - "INFO:agent.ComputerAgent:Total usage:\n", - " - completion_tokens: 1266\n", - " - prompt_tokens: 12578\n", - " - total_tokens: 13844\n", - " - completion_tokens_details:\n", - " - accepted_prediction_tokens: 0\n", - " - audio_tokens: 0\n", - " - reasoning_tokens: 1024\n", - " - rejected_prediction_tokens: 0\n", - " - prompt_tokens_details:\n", - " - audio_tokens: 0\n", - " - cached_tokens: 11776\n", - " - response_cost: $0.0151\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:29:26 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Loading checkpoint shards: 50%|█████ | 2/4 [00:03<00:03, 1.67s/it]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:29:27 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m19:29:28 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 400 Bad Request\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 91%|████████████████████████████████████----| 6708/7340 [243:09<22:54, 27.6 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Loading checkpoint shards: 75%|███████▌ | 3/4 [00:04<00:01, 1.64s/it]2025-08-11 19:29:29,309 - agent.ComputerAgent - INFO - Computer: type({'text': 'sar -V\\n'})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': 'sar -V\\n'})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Loading checkpoint shards: 100%|██████████| 4/4 [00:05<00:00, 1.38s/it]\n", - "\u001b[92m19:29:30 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 91%|████████████████████████████████████----| 6709/7340 [243:11<22:52, 27.6 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m19:29:31 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 91%|████████████████████████████████████----| 6710/7340 [243:13<22:50, 27.6 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m19:29:31 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/73c70c0d-c1a0-401f-83c0-063e983abd6c/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/6f055f32-c606-4a58-91bc-c5990d4e8280/invoke \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:29:32,337 - agent.ComputerAgent - INFO - LLM processing started with 29 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 29 messages\n", - "\u001b[92m19:29:32 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:29:33,023 - agent.ComputerAgent - INFO - Computer: click({'x': 115, 'y': 635})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 115, 'y': 635})\n", - "\u001b[92m19:29:33 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 91%|████████████████████████████████████----| 6710/7340 [243:14<22:50, 27.6 steps/min]\u001b[92m19:29:33 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:29:33,677 - agent.ComputerAgent - INFO - Computer: click({'x': 28, 'y': 739})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 28, 'y': 739})\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:29:34,335 - agent.ComputerAgent - INFO - Computer: click({'x': 530, 'y': 417})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 530, 'y': 417})\n", - "\u001b[92m19:29:34 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m19:29:34 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:29:34,988 - agent.ComputerAgent - INFO - Agent: Taking a screenshot to see the current computer screen.\n", - "INFO:agent.ComputerAgent:Agent: Taking a screenshot to see the current computer screen.\n", - "2025-08-11 19:29:34,989 - agent.ComputerAgent - INFO - Computer: click({'x': 80, 'y': 181})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 80, 'y': 181})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - " 91%|████████████████████████████████████----| 6711/7340 [243:17<22:48, 27.6 steps/min]\u001b[92m19:29:35 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m19:29:35 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/c915dbd9-32bc-40a7-9c07-d437c737419f/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m19:29:35 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m19:29:35 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/775a5b67-2406-42b8-86e5-243e01b8dc27/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/18debd9e-6c58-4504-8a04-13cba683a254/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/7a2c2835-d21e-4e04-babb-e8305a4f1f9d/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/77892268-14f2-4dfa-b58c-6a682f258679/invoke \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:29:36,366 - agent.ComputerAgent - INFO - Computer: click({'x': 186, 'y': 148})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 186, 'y': 148})\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:29:36,994 - agent.ComputerAgent - INFO - Computer: click({'x': 85, 'y': 234})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 85, 'y': 234})\n", - "\u001b[92m19:29:37 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m19:29:37 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 91%|████████████████████████████████████----| 6714/7340 [243:18<22:41, 27.6 steps/min]2025-08-11 19:29:37,669 - agent.ComputerAgent - INFO - Computer: click({'x': 483, 'y': 267})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 483, 'y': 267})\n", - "2025-08-11 19:29:38,353 - agent.ComputerAgent - INFO - Computer: click({'x': 974, 'y': 34})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 974, 'y': 34})\n", - "2025-08-11 19:29:38,999 - agent.ComputerAgent - INFO - LLM processing started with 22 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 22 messages\n", - "\u001b[92m19:29:39 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m19:29:39 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 400 Bad Request\"\n", - " 92%|████████████████████████████████████----| 6717/7340 [243:20<22:34, 27.6 steps/min]2025-08-11 19:29:39,730 - agent.ComputerAgent - INFO - Computer: drag({'path': [{'x': 914, 'y': 671}, {'x': 984, 'y': 467}]})\n", - "INFO:agent.ComputerAgent:Computer: drag({'path': [{'x': 914, 'y': 671}, {'x': 984, 'y': 467}]})\n", - "2025-08-11 19:29:40,390 - agent.ComputerAgent - INFO - LLM processing started with 42 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 42 messages\n", - "\u001b[92m19:29:40 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 92%|████████████████████████████████████----| 6719/7340 [243:22<22:29, 27.6 steps/min]2025-08-11 19:29:41,077 - agent.ComputerAgent - INFO - LLM processing started with 18 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 18 messages\n", - "\u001b[92m19:29:41 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 19:29:42,161 - agent.ComputerAgent - INFO - LLM processing started with 22 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 22 messages\n", - "\u001b[92m19:29:42 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/73c70c0d-c1a0-401f-83c0-063e983abd6c/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/6f055f32-c606-4a58-91bc-c5990d4e8280/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/18debd9e-6c58-4504-8a04-13cba683a254/close \"HTTP/1.1 200 OK\"\n", - " 92%|████████████████████████████████████----| 6720/7340 [243:23<22:27, 27.6 steps/min]2025-08-11 19:29:42,846 - agent.ComputerAgent - INFO - LLM processing started with 31 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 31 messages\n", - "\u001b[92m19:29:42 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 92%|████████████████████████████████████----| 6725/7340 [243:26<22:15, 27.6 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:29:45 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/0180c5d2-a012-4261-b093-ed34f443f269/invoke \"HTTP/1.1 200 OK\"\n", - " 92%|████████████████████████████████████----| 6725/7340 [243:27<22:15, 27.6 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/052ac585-1998-46b2-9ac5-0dc192aeba02/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/6f055f32-c606-4a58-91bc-c5990d4e8280/close \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/963f0b0a-47d1-479c-9077-6c59023108fe/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/84a5d283-63f1-43fc-b483-76116d67f385/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/1a178f89-87e5-46d9-a114-22d5fcc5c630/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/f209659e-2ed9-40fa-ae61-5359fb9ef290/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/bcec4523-df7a-48b5-aea1-8d7c632a6dc4/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/9a544504-3e48-48b2-8429-0a97e266ebfb/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/af58ffed-65a3-4c4a-a9fe-5c940230627d/invoke \"HTTP/1.1 200 OK\"\n", - " 92%|████████████████████████████████████----| 6725/7340 [243:28<22:15, 27.6 steps/min]2025-08-11 19:29:48,280 - agent.ComputerAgent - INFO - LLM processing started with 6 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 6 messages\n", - "\u001b[92m19:29:48 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "Loading checkpoint shards: 0%| | 0/4 [00:00 System_Resources_Report.txt\\n'})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': 'sar -u 1 30 > System_Resources_Report.txt\\n'})\n", - " 92%|████████████████████████████████████----| 6784/7340 [245:45<20:08, 27.6 steps/min]2025-08-11 19:32:03,959 - agent.ComputerAgent - INFO - LLM processing started with 6 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 6 messages\n", - "\u001b[92m19:32:03 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 19:32:04,639 - agent.ComputerAgent - INFO - LLM processing started with 30 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 30 messages\n", - "\u001b[92m19:32:04 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/c915dbd9-32bc-40a7-9c07-d437c737419f/invoke \"HTTP/1.1 200 OK\"\n", - " 92%|████████████████████████████████████----| 6785/7340 [245:46<20:06, 27.6 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/7633715b-dde0-4c56-a4b6-22ccee78f5f5/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/963f0b0a-47d1-479c-9077-6c59023108fe/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:32:05,820 - agent.ComputerAgent - INFO - LLM processing started with 18 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 18 messages\n", - "\u001b[92m19:32:05 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v1/environments/6f3b006b-141d-439d-b6cb-eed7bd6483c3/reset \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/7f112db6-0b60-4e6c-86f5-0d87dc91f371/invoke \"HTTP/1.1 200 OK\"\n", - " 92%|████████████████████████████████████----| 6785/7340 [245:47<20:06, 27.6 steps/min]2025-08-11 19:32:06,520 - agent.ComputerAgent - INFO - LLM processing started with 12 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 12 messages\n", - "\u001b[92m19:32:06 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 92%|████████████████████████████████████----| 6785/7340 [245:48<20:06, 27.6 steps/min]2025-08-11 19:32:07,704 - agent.ComputerAgent - INFO - LLM processing started with 28 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 28 messages\n", - "\u001b[92m19:32:07 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 19:32:08,380 - agent.ComputerAgent - INFO - LLM processing started with 8 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 8 messages\n", - "\u001b[92m19:32:08 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 92%|████████████████████████████████████----| 6785/7340 [245:50<20:06, 27.6 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/51954fb4-34ed-4511-b2fd-a6169b5ea5d3/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:32:10,389 - agent.ComputerAgent - INFO - LLM processing started with 42 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 42 messages\n", - "\u001b[92m19:32:10 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/775a5b67-2406-42b8-86e5-243e01b8dc27/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:32:11 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/0c64a3b4-e9b0-46c1-a580-cdcf62b74e44/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 92%|████████████████████████████████████----| 6785/7340 [245:53<20:06, 27.6 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m19:32:12 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/6f3b006b-141d-439d-b6cb-eed7bd6483c3/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:32:12,779 - agent.ComputerAgent - INFO - LLM processing started with 1 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 1 messages\n", - "\u001b[92m19:32:12 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m19:32:12 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 92%|████████████████████████████████████----| 6785/7340 [245:54<20:06, 27.6 steps/min]2025-08-11 19:32:13,458 - agent.ComputerAgent - INFO - Computer: click({'x': 461, 'y': 169})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 461, 'y': 169})\n", - "\u001b[92m19:32:13 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:32:14,530 - agent.ComputerAgent - INFO - Computer: click({'x': 125, 'y': 182})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 125, 'y': 182})\n", - " 92%|████████████████████████████████████----| 6785/7340 [245:56<20:07, 27.6 steps/min]2025-08-11 19:32:15,198 - agent.ComputerAgent - INFO - LLM processing started with 8 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 8 messages\n", - "\u001b[92m19:32:15 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:32:16,538 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'win'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'win'})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - " 92%|████████████████████████████████████----| 6787/7340 [245:58<20:02, 27.6 steps/min]\u001b[92m19:32:17 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:32:17 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:32:19,209 - agent.ComputerAgent - INFO - Computer: type({'text': 'Dublin'})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': 'Dublin'})\n", - " 92%|████████████████████████████████████----| 6788/7340 [246:00<20:00, 27.6 steps/min]\u001b[92m19:32:19 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:32:19,877 - agent.ComputerAgent - INFO - LLM processing started with 26 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 26 messages\n", - "\u001b[92m19:32:19 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 19:32:20,569 - agent.ComputerAgent - INFO - Computer: click({'x': 90, 'y': 183})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 90, 'y': 183})\n", - "\u001b[92m19:32:20 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:32:21,243 - agent.ComputerAgent - INFO - Computer: click({'x': 430, 'y': 219})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 430, 'y': 219})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - " 92%|████████████████████████████████████----| 6789/7340 [246:04<19:58, 27.6 steps/min]\u001b[92m19:32:22 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m19:32:22 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m19:32:23 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:32:23,644 - agent.ComputerAgent - INFO - Computer: click({'x': 188, 'y': 190})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 188, 'y': 190})\n", - " 93%|█████████████████████████████████████---| 6791/7340 [246:05<19:53, 27.6 steps/min]\u001b[92m19:32:24 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:32:24,280 - agent.ComputerAgent - INFO - Computer: click({'x': 123, 'y': 178})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 123, 'y': 178})\n", - " 93%|█████████████████████████████████████---| 6793/7340 [246:07<19:49, 27.6 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:32:26 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - " 93%|█████████████████████████████████████---| 6793/7340 [246:08<19:49, 27.6 steps/min]\u001b[92m19:32:27 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:32:27,608 - agent.ComputerAgent - INFO - Computer: click({'x': 18, 'y': 476})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 18, 'y': 476})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/b928bd01-f1b7-4f34-accf-acb6aec5d8cd/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/7a2c2835-d21e-4e04-babb-e8305a4f1f9d/invoke \"HTTP/1.1 200 OK\"\n", - " 93%|█████████████████████████████████████---| 6793/7340 [246:09<19:49, 27.6 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/0180c5d2-a012-4261-b093-ed34f443f269/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:32:28,295 - agent.ComputerAgent - INFO - LLM processing started with 8 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 8 messages\n", - "\u001b[92m19:32:28 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/7633715b-dde0-4c56-a4b6-22ccee78f5f5/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 19:32:28,953 - agent.ComputerAgent - INFO - LLM processing started with 30 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 30 messages\n", - "\u001b[92m19:32:28 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v1/environments/acf3037a-4b6c-4ea8-b81c-ffc2e76132e1/reset \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/1a178f89-87e5-46d9-a114-22d5fcc5c630/invoke \"HTTP/1.1 200 OK\"\n", - " 93%|█████████████████████████████████████---| 6794/7340 [246:10<19:47, 27.6 steps/min]2025-08-11 19:32:29,623 - agent.ComputerAgent - INFO - LLM processing started with 14 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 14 messages\n", - "\u001b[92m19:32:29 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:32:30 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/963f0b0a-47d1-479c-9077-6c59023108fe/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/f209659e-2ed9-40fa-ae61-5359fb9ef290/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/7f112db6-0b60-4e6c-86f5-0d87dc91f371/invoke \"HTTP/1.1 200 OK\"\n", - " 93%|█████████████████████████████████████---| 6794/7340 [246:12<19:47, 27.6 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:32:30,976 - agent.ComputerAgent - INFO - LLM processing started with 20 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 20 messages\n", - "\u001b[92m19:32:30 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 19:32:31,628 - agent.ComputerAgent - INFO - LLM processing started with 14 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 14 messages\n", - "\u001b[92m19:32:31 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m19:32:31 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 93%|█████████████████████████████████████---| 6794/7340 [246:13<19:47, 27.6 steps/min]2025-08-11 19:32:32,293 - agent.ComputerAgent - INFO - Computer: click({'x': 534, 'y': 554})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 534, 'y': 554})\n", - "2025-08-11 19:32:32,937 - agent.ComputerAgent - INFO - LLM processing started with 10 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 10 messages\n", - "\u001b[92m19:32:32 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 93%|█████████████████████████████████████---| 6794/7340 [246:14<19:47, 27.6 steps/min]2025-08-11 19:32:33,607 - agent.ComputerAgent - INFO - LLM processing started with 28 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 28 messages\n", - "\u001b[92m19:32:33 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 19:32:34,249 - agent.ComputerAgent - INFO - LLM processing started with 32 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 32 messages\n", - "\u001b[92m19:32:34 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 93%|█████████████████████████████████████---| 6795/7340 [246:17<19:45, 27.6 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/0c64a3b4-e9b0-46c1-a580-cdcf62b74e44/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:32:36,916 - agent.ComputerAgent - INFO - LLM processing started with 10 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 10 messages\n", - "\u001b[92m19:32:36 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/create_environment \"HTTP/1.1 200 OK\"\n", - " 93%|█████████████████████████████████████---| 6795/7340 [246:18<19:45, 27.6 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:32:37 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/acf3037a-4b6c-4ea8-b81c-ffc2e76132e1/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - " 93%|█████████████████████████████████████---| 6795/7340 [246:19<19:45, 27.6 steps/min]2025-08-11 19:32:38,805 - agent.ComputerAgent - INFO - LLM processing started with 1 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 1 messages\n", - "\u001b[92m19:32:38 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/51c56274-d8ae-4edf-8ff1-b637cd2fff66/invoke \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:32:38 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:32:39,495 - agent.ComputerAgent - INFO - Computer: click({'x': 173, 'y': 149})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 173, 'y': 149})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/af58ffed-65a3-4c4a-a9fe-5c940230627d/invoke \"HTTP/1.1 200 OK\"\n", - " 93%|█████████████████████████████████████---| 6795/7340 [246:21<19:45, 27.6 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:32:40,803 - agent.ComputerAgent - INFO - Computer: type({'text': 'wc -l System_Resources_Report.txt\\n'})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': 'wc -l System_Resources_Report.txt\\n'})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:32:41 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 93%|█████████████████████████████████████---| 6796/7340 [246:23<19:43, 27.6 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:32:42,099 - agent.ComputerAgent - INFO - LLM processing started with 24 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 24 messages\n", - "\u001b[92m19:32:42 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 93%|█████████████████████████████████████---| 6797/7340 [246:24<19:41, 27.6 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:32:43 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:32:43 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m19:32:43 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:32:44,597 - agent.ComputerAgent - INFO - Computer: double_click({'x': 482, 'y': 277})\n", - "INFO:agent.ComputerAgent:Computer: double_click({'x': 482, 'y': 277})\n", - " 93%|█████████████████████████████████████---| 6797/7340 [246:26<19:41, 27.6 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m19:32:45 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:32:45 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:32:46,644 - agent.ComputerAgent - INFO - Computer: type({'text': '=SUM(Sheet1.B2:B11)'})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': '=SUM(Sheet1.B2:B11)'})\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:32:47,299 - agent.ComputerAgent - INFO - Agent: Taking a screenshot to see the current computer screen.\n", - "INFO:agent.ComputerAgent:Agent: Taking a screenshot to see the current computer screen.\n", - "2025-08-11 19:32:47,300 - agent.ComputerAgent - INFO - Computer: click({'x': 578, 'y': 286})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 578, 'y': 286})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:32:47 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 93%|█████████████████████████████████████---| 6798/7340 [246:29<19:39, 27.6 steps/min]\u001b[92m19:32:48 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:32:48,654 - agent.ComputerAgent - INFO - Computer: click({'x': 316, 'y': 416})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 316, 'y': 416})\n", - "\u001b[92m19:32:48 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/775a5b67-2406-42b8-86e5-243e01b8dc27/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/c915dbd9-32bc-40a7-9c07-d437c737419f/invoke \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:32:49,313 - agent.ComputerAgent - INFO - Computer: click({'x': 306, 'y': 416})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 306, 'y': 416})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:32:50 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m19:32:50 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 93%|█████████████████████████████████████---| 6800/7340 [246:31<19:34, 27.6 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:32:50,681 - agent.ComputerAgent - INFO - Computer: click({'x': 237, 'y': 254})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 237, 'y': 254})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:32:51 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:32:51 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m19:32:52 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 93%|█████████████████████████████████████---| 6802/7340 [246:33<19:30, 27.6 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:32:52,640 - agent.ComputerAgent - INFO - Agent: Taking a screenshot to see the current computer screen.\n", - "INFO:agent.ComputerAgent:Agent: Taking a screenshot to see the current computer screen.\n", - "2025-08-11 19:32:52,640 - agent.ComputerAgent - INFO - Computer: click({'x': 14, 'y': 524})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 14, 'y': 524})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:32:54,035 - agent.ComputerAgent - INFO - Computer: type({'text': '=A2/1000000'})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': '=A2/1000000'})\n", - "\u001b[92m19:32:54 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m19:32:54 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 93%|█████████████████████████████████████---| 6803/7340 [246:35<19:27, 27.6 steps/min]2025-08-11 19:32:54,689 - agent.ComputerAgent - INFO - Computer: click({'x': 19, 'y': 481})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 19, 'y': 481})\n", - "2025-08-11 19:32:55,406 - agent.ComputerAgent - INFO - Computer: click({'x': 237, 'y': 193})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 237, 'y': 193})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:32:56 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 93%|█████████████████████████████████████---| 6805/7340 [246:37<19:23, 27.6 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:32:56,702 - agent.ComputerAgent - INFO - LLM processing started with 28 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 28 messages\n", - "\u001b[92m19:32:56 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 19:32:57,376 - agent.ComputerAgent - INFO - LLM processing started with 30 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 30 messages\n", - "\u001b[92m19:32:57 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m19:32:57 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/6f3b006b-141d-439d-b6cb-eed7bd6483c3/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/963f0b0a-47d1-479c-9077-6c59023108fe/invoke \"HTTP/1.1 200 OK\"\n", - " 93%|█████████████████████████████████████---| 6807/7340 [246:39<19:18, 27.6 steps/min]2025-08-11 19:32:58,080 - agent.ComputerAgent - INFO - Computer: click({'x': 343, 'y': 183})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 343, 'y': 183})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/b928bd01-f1b7-4f34-accf-acb6aec5d8cd/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:32:58,735 - agent.ComputerAgent - INFO - LLM processing started with 22 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 22 messages\n", - "\u001b[92m19:32:58 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 19:32:59,441 - agent.ComputerAgent - INFO - LLM processing started with 6 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 6 messages\n", - "\u001b[92m19:32:59 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 93%|█████████████████████████████████████---| 6807/7340 [246:41<19:18, 27.6 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:33:00 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:33:00 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/7633715b-dde0-4c56-a4b6-22ccee78f5f5/invoke \"HTTP/1.1 200 OK\"\n", - " 93%|█████████████████████████████████████---| 6808/7340 [246:42<19:16, 27.6 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:33:01,450 - agent.ComputerAgent - INFO - LLM processing started with 10 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 10 messages\n", - "\u001b[92m19:33:01 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/7f112db6-0b60-4e6c-86f5-0d87dc91f371/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/7a2c2835-d21e-4e04-babb-e8305a4f1f9d/invoke \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:33:02,154 - agent.ComputerAgent - INFO - LLM processing started with 16 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 16 messages\n", - "\u001b[92m19:33:02 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m19:33:02 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/0c64a3b4-e9b0-46c1-a580-cdcf62b74e44/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 93%|█████████████████████████████████████---| 6808/7340 [246:43<19:16, 27.6 steps/min]2025-08-11 19:33:02,855 - agent.ComputerAgent - INFO - LLM processing started with 32 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 32 messages\n", - "\u001b[92m19:33:02 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 19:33:03,530 - agent.ComputerAgent - INFO - Computer: click({'x': 633, 'y': 473})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 633, 'y': 473})\n", - "\u001b[92m19:33:03 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/acf3037a-4b6c-4ea8-b81c-ffc2e76132e1/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/51954fb4-34ed-4511-b2fd-a6169b5ea5d3/invoke \"HTTP/1.1 200 OK\"\n", - " 93%|█████████████████████████████████████---| 6808/7340 [246:45<19:16, 27.6 steps/min]\u001b[92m19:33:03 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:33:04,220 - agent.ComputerAgent - INFO - LLM processing started with 12 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 12 messages\n", - "\u001b[92m19:33:04 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/f209659e-2ed9-40fa-ae61-5359fb9ef290/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:33:04,896 - agent.ComputerAgent - INFO - LLM processing started with 12 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 12 messages\n", - "\u001b[92m19:33:04 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m19:33:04 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 93%|█████████████████████████████████████---| 6809/7340 [246:46<19:14, 27.6 steps/min]2025-08-11 19:33:05,586 - agent.ComputerAgent - INFO - Computer: drag({'path': [{'x': 424, 'y': 418}, {'x': 527, 'y': 226}]})\n", - "INFO:agent.ComputerAgent:Computer: drag({'path': [{'x': 424, 'y': 418}, {'x': 527, 'y': 226}]})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:33:06 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/0180c5d2-a012-4261-b093-ed34f443f269/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 93%|█████████████████████████████████████---| 6809/7340 [246:48<19:14, 27.6 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:33:06,917 - agent.ComputerAgent - INFO - LLM processing started with 16 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 16 messages\n", - "\u001b[92m19:33:06 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 19:33:07,597 - agent.ComputerAgent - INFO - LLM processing started with 6 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 6 messages\n", - "\u001b[92m19:33:07 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m19:33:07 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 93%|█████████████████████████████████████---| 6810/7340 [246:49<19:12, 27.6 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:33:08,272 - agent.ComputerAgent - INFO - Computer: click({'x': 946, 'y': 750})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 946, 'y': 750})\n", - "2025-08-11 19:33:08,957 - agent.ComputerAgent - INFO - LLM processing started with 34 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 34 messages\n", - "\u001b[92m19:33:09 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 93%|█████████████████████████████████████---| 6811/7340 [246:51<19:10, 27.6 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/51954fb4-34ed-4511-b2fd-a6169b5ea5d3/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/af58ffed-65a3-4c4a-a9fe-5c940230627d/invoke \"HTTP/1.1 200 OK\"\n", - " 93%|█████████████████████████████████████---| 6812/7340 [246:52<19:08, 27.6 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/51954fb4-34ed-4511-b2fd-a6169b5ea5d3/close \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:33:13,491 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'enter'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'enter'})\n", - " 93%|█████████████████████████████████████---| 6812/7340 [246:55<19:08, 27.6 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:33:14,808 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'enter'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'enter'})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:33:15 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m19:33:16 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/b928bd01-f1b7-4f34-accf-acb6aec5d8cd/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/1a178f89-87e5-46d9-a114-22d5fcc5c630/invoke \"HTTP/1.1 200 OK\"\n", - " 93%|█████████████████████████████████████---| 6813/7340 [246:57<19:06, 27.6 steps/min]2025-08-11 19:33:16,837 - agent.ComputerAgent - INFO - LLM processing started with 26 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 26 messages\n", - "\u001b[92m19:33:16 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 19:33:17,526 - agent.ComputerAgent - INFO - LLM processing started with 12 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 12 messages\n", - "\u001b[92m19:33:17 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 93%|█████████████████████████████████████---| 6814/7340 [246:59<19:03, 27.6 steps/min]2025-08-11 19:33:18,187 - agent.ComputerAgent - INFO - LLM processing started with 30 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 30 messages\n", - "\u001b[92m19:33:18 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: GET https://orchestration.hud.so/hud-gym/api/v1/gyms/OSWorld-Ubuntu \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 93%|█████████████████████████████████████---| 6814/7340 [247:00<19:04, 27.6 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:33:20,027 - agent.ComputerAgent - INFO - Computer: click({'x': 89, 'y': 185})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 89, 'y': 185})\n", - " 93%|█████████████████████████████████████---| 6814/7340 [247:01<19:04, 27.6 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:33:21,352 - agent.ComputerAgent - INFO - Computer: type({'text': 'find / -type f -name \"secret.docx\" 2>/dev/null'})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': 'find / -type f -name \"secret.docx\" 2>/dev/null'})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:33:22 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 93%|█████████████████████████████████████---| 6815/7340 [247:03<19:01, 27.6 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/7a2c2835-d21e-4e04-babb-e8305a4f1f9d/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/963f0b0a-47d1-479c-9077-6c59023108fe/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:33:22,639 - agent.ComputerAgent - INFO - LLM processing started with 34 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 34 messages\n", - "\u001b[92m19:33:22 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 19:33:23,336 - agent.ComputerAgent - INFO - LLM processing started with 24 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 24 messages\n", - "\u001b[92m19:33:23 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 93%|█████████████████████████████████████---| 6816/7340 [247:05<18:59, 27.6 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Loading checkpoint shards: 0%| | 0/4 [00:00/dev/null | head -n1); printf \"%s\\\\n\" \"$path\"; echo -n \"$path\" | xclip -selection clipboard 2>/dev/null || echo -n \"$path\" | wl-copy 2>/dev/null || true'})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': 'path=$(find / -type f -name \"secret.docx\" 2>/dev/null | head -n1); printf \"%s\\\\n\" \"$path\"; echo -n \"$path\" | xclip -selection clipboard 2>/dev/null || echo -n \"$path\" | wl-copy 2>/dev/null || true'})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Loading checkpoint shards: 50%|█████ | 2/4 [00:03<00:03, 1.76s/it]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:37:48 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Loading checkpoint shards: 75%|███████▌ | 3/4 [00:05<00:01, 1.72s/it]2025-08-11 19:37:50,401 - agent.ComputerAgent - INFO - Agent: The video has been converted to an MP3 and saved on the Desktop as “Baby Justin Bieber.mp3.” You can now play it anytime.\n", - "\n", - "Task completed\n", - "INFO:agent.ComputerAgent:Agent: The video has been converted to an MP3 and saved on the Desktop as “Baby Justin Bieber.mp3.” You can now play it anytime.\n", - "\n", - "Task completed\n", - "Loading checkpoint shards: 100%|██████████| 4/4 [00:05<00:00, 1.45s/it]\n", - "2025-08-11 19:37:51,060 - agent.ComputerAgent - INFO - Total usage:\n", - " - completion_tokens: 488\n", - " - prompt_tokens: 12887\n", - " - total_tokens: 13375\n", - " - completion_tokens_details:\n", - " - accepted_prediction_tokens: 0\n", - " - audio_tokens: 0\n", - " - reasoning_tokens: 448\n", - " - rejected_prediction_tokens: 0\n", - " - prompt_tokens_details:\n", - " - audio_tokens: 0\n", - " - cached_tokens: 0\n", - " - response_cost: $0.0210\n", - "INFO:agent.ComputerAgent:Total usage:\n", - " - completion_tokens: 488\n", - " - prompt_tokens: 12887\n", - " - total_tokens: 13375\n", - " - completion_tokens_details:\n", - " - accepted_prediction_tokens: 0\n", - " - audio_tokens: 0\n", - " - reasoning_tokens: 448\n", - " - rejected_prediction_tokens: 0\n", - " - prompt_tokens_details:\n", - " - audio_tokens: 0\n", - " - cached_tokens: 0\n", - " - response_cost: $0.0210\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:37:51 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m19:37:52 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 95%|█████████████████████████████████████---| 6944/7340 [251:34<14:20, 27.6 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m19:37:53 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m19:37:53 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:37:53,761 - agent.ComputerAgent - INFO - Computer: click({'x': 515, 'y': 232})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 515, 'y': 232})\n", - "\u001b[92m19:37:53 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 95%|█████████████████████████████████████---| 6945/7340 [251:35<14:18, 27.6 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:37:54,423 - agent.ComputerAgent - INFO - Computer: scroll({'scroll_y': 640, 'scroll_x': 0, 'x': 989, 'y': 366})\n", - "INFO:agent.ComputerAgent:Computer: scroll({'scroll_y': 640, 'scroll_x': 0, 'x': 989, 'y': 366})\n", - "\u001b[92m19:37:54 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:37:55,120 - agent.ComputerAgent - INFO - Computer: double_click({'x': 94, 'y': 453})\n", - "INFO:agent.ComputerAgent:Computer: double_click({'x': 94, 'y': 453})\n", - "\u001b[92m19:37:55 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 95%|█████████████████████████████████████---| 6946/7340 [251:36<14:16, 27.6 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:37:55,799 - agent.ComputerAgent - INFO - Computer: click({'x': 153, 'y': 52})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 153, 'y': 52})\n", - "\u001b[92m19:37:55 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:37:56,468 - agent.ComputerAgent - INFO - Computer: double_click({'x': 354, 'y': 136})\n", - "INFO:agent.ComputerAgent:Computer: double_click({'x': 354, 'y': 136})\n", - "\u001b[92m19:37:56 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m19:37:57 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 95%|█████████████████████████████████████---| 6948/7340 [251:38<14:11, 27.6 steps/min]2025-08-11 19:37:57,837 - agent.ComputerAgent - INFO - Computer: click({'x': 471, 'y': 205})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 471, 'y': 205})\n", - "\u001b[92m19:37:57 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/af58ffed-65a3-4c4a-a9fe-5c940230627d/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m19:37:57 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:37:58,540 - agent.ComputerAgent - INFO - Computer: click({'x': 945, 'y': 500})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 945, 'y': 500})\n", - "2025-08-11 19:37:59,244 - agent.ComputerAgent - INFO - Computer: double_click({'x': 989, 'y': 713})\n", - "INFO:agent.ComputerAgent:Computer: double_click({'x': 989, 'y': 713})\n", - " 95%|█████████████████████████████████████---| 6950/7340 [251:40<14:07, 27.6 steps/min]\u001b[92m19:37:59 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:37:59,918 - agent.ComputerAgent - INFO - Computer: click({'x': 913, 'y': 184})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 913, 'y': 184})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/0180c5d2-a012-4261-b093-ed34f443f269/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:38:00,585 - agent.ComputerAgent - INFO - LLM processing started with 32 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 32 messages\n", - "\u001b[92m19:38:00 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 95%|█████████████████████████████████████---| 6953/7340 [251:42<14:00, 27.6 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/acf3037a-4b6c-4ea8-b81c-ffc2e76132e1/invoke \"HTTP/1.1 200 OK\"\n", - " 95%|█████████████████████████████████████---| 6954/7340 [251:43<13:58, 27.6 steps/min]2025-08-11 19:38:02,285 - agent.ComputerAgent - INFO - LLM processing started with 22 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 22 messages\n", - "\u001b[92m19:38:02 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 95%|█████████████████████████████████████---| 6954/7340 [251:44<13:58, 27.6 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/7f112db6-0b60-4e6c-86f5-0d87dc91f371/invoke \"HTTP/1.1 200 OK\"\n", - " 95%|█████████████████████████████████████---| 6954/7340 [251:45<13:58, 27.6 steps/min]2025-08-11 19:38:04,439 - agent.ComputerAgent - INFO - LLM processing started with 28 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 28 messages\n", - "\u001b[92m19:38:04 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/51c56274-d8ae-4edf-8ff1-b637cd2fff66/invoke \"HTTP/1.1 200 OK\"\n", - " 95%|█████████████████████████████████████---| 6954/7340 [251:46<13:58, 27.6 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/8e75deb1-3c97-408b-8c7d-f4681b322141/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:38:05,613 - agent.ComputerAgent - INFO - LLM processing started with 10 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 10 messages\n", - "\u001b[92m19:38:05 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/c915dbd9-32bc-40a7-9c07-d437c737419f/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:38:07,021 - agent.ComputerAgent - INFO - Agent: I configured your system to lock automatically when you’re away.\n", - "\n", - "What I set\n", - "- Enabled Automatic Screen Lock.\n", - "- Set Blank Screen Delay to 5 minutes, so after 5 minutes of inactivity the screen turns off and locks immediately.\n", - "- Kept “Lock screen on Suspend” enabled, so any suspend will also require a password on wake.\n", - "\n", - "How to adjust later\n", - "- Settings > Privacy > Screen:\n", - " - Change “Blank Screen Delay” to your preferred timeout (e.g., 1–2 minutes for tighter security).\n", - " - “Automatic Screen Lock Delay” is set to lock as soon as the screen turns off.\n", - "\n", - "Tip\n", - "- Press Super+L anytime to lock instantly.\n", - "\n", - "Task completed.\n", - "INFO:agent.ComputerAgent:Agent: I configured your system to lock automatically when you’re away.\n", - "\n", - "What I set\n", - "- Enabled Automatic Screen Lock.\n", - "- Set Blank Screen Delay to 5 minutes, so after 5 minutes of inactivity the screen turns off and locks immediately.\n", - "- Kept “Lock screen on Suspend” enabled, so any suspend will also require a password on wake.\n", - "\n", - "How to adjust later\n", - "- Settings > Privacy > Screen:\n", - " - Change “Blank Screen Delay” to your preferred timeout (e.g., 1–2 minutes for tighter security).\n", - " - “Automatic Screen Lock Delay” is set to lock as soon as the screen turns off.\n", - "\n", - "Tip\n", - "- Press Super+L anytime to lock instantly.\n", - "\n", - "Task completed.\n", - "2025-08-11 19:38:07,641 - agent.ComputerAgent - INFO - Total usage:\n", - " - completion_tokens: 790\n", - " - prompt_tokens: 10768\n", - " - total_tokens: 11558\n", - " - completion_tokens_details:\n", - " - accepted_prediction_tokens: 0\n", - " - audio_tokens: 0\n", - " - reasoning_tokens: 640\n", - " - rejected_prediction_tokens: 0\n", - " - prompt_tokens_details:\n", - " - audio_tokens: 0\n", - " - cached_tokens: 0\n", - " - response_cost: $0.0214\n", - "INFO:agent.ComputerAgent:Total usage:\n", - " - completion_tokens: 790\n", - " - prompt_tokens: 10768\n", - " - total_tokens: 11558\n", - " - completion_tokens_details:\n", - " - accepted_prediction_tokens: 0\n", - " - audio_tokens: 0\n", - " - reasoning_tokens: 640\n", - " - rejected_prediction_tokens: 0\n", - " - prompt_tokens_details:\n", - " - audio_tokens: 0\n", - " - cached_tokens: 0\n", - " - response_cost: $0.0214\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/5e73167c-1836-4752-b7e8-57434e5d7875/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/ba3f752c-7268-49e8-819c-5b471e52bd54/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/0c64a3b4-e9b0-46c1-a580-cdcf62b74e44/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/25f45afe-ee57-4629-9991-c515438accab/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/68f513cf-ec21-4216-bab9-84c5bfcfce88/invoke \"HTTP/1.1 200 OK\"\n", - " 95%|█████████████████████████████████████---| 6955/7340 [251:49<13:56, 27.6 steps/min]2025-08-11 19:38:08,315 - agent.ComputerAgent - INFO - LLM processing started with 16 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 16 messages\n", - "\u001b[92m19:38:08 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:38:09 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:38:09,672 - agent.ComputerAgent - INFO - LLM processing started with 14 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 14 messages\n", - "\u001b[92m19:38:09 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - " 95%|█████████████████████████████████████---| 6955/7340 [251:52<13:56, 27.6 steps/min]\u001b[92m19:38:10 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m19:38:10 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:38:11,032 - agent.ComputerAgent - INFO - Computer: click({'x': 375, 'y': 75})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 375, 'y': 75})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/af58ffed-65a3-4c4a-a9fe-5c940230627d/invoke \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:38:11 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:38:11,731 - agent.ComputerAgent - INFO - Computer: click({'x': 210, 'y': 202})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 210, 'y': 202})\n", - " 95%|█████████████████████████████████████---| 6958/7340 [251:53<13:49, 27.6 steps/min]2025-08-11 19:38:12,415 - agent.ComputerAgent - INFO - LLM processing started with 30 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 30 messages\n", - "\u001b[92m19:38:12 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/7633715b-dde0-4c56-a4b6-22ccee78f5f5/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:38:13,115 - agent.ComputerAgent - INFO - LLM processing started with 14 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 14 messages\n", - "\u001b[92m19:38:13 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 95%|█████████████████████████████████████---| 6960/7340 [251:54<13:45, 27.6 steps/min]2025-08-11 19:38:13,805 - agent.ComputerAgent - INFO - LLM processing started with 40 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 40 messages\n", - "\u001b[92m19:38:13 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/af58ffed-65a3-4c4a-a9fe-5c940230627d/close \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:38:14,486 - agent.ComputerAgent - INFO - LLM processing started with 10 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 10 messages\n", - "\u001b[92m19:38:14 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 19:38:15,845 - agent.ComputerAgent - INFO - LLM processing started with 18 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 18 messages\n", - "\u001b[92m19:38:15 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 95%|█████████████████████████████████████---| 6960/7340 [251:57<13:45, 27.6 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:38:17,724 - agent.ComputerAgent - INFO - Computer: type({'text': 'Vienna'})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': 'Vienna'})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/7633715b-dde0-4c56-a4b6-22ccee78f5f5/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:38:19,084 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'ENTER'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'ENTER'})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v1/environments/40c5f987-3d81-47fe-8798-4e45d9755f93/reset \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: GET https://orchestration.hud.so/hud-gym/api/v1/gyms/OSWorld-Ubuntu \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/6f3b006b-141d-439d-b6cb-eed7bd6483c3/invoke \"HTTP/1.1 200 OK\"\n", - " 95%|█████████████████████████████████████---| 6960/7340 [252:00<13:45, 27.6 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:38:20,825 - agent.ComputerAgent - INFO - Computer: click({'x': 342, 'y': 184})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 342, 'y': 184})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/31367309-0055-409a-a992-edf729fb010c/invoke \"HTTP/1.1 200 OK\"\n", - " 95%|█████████████████████████████████████---| 6969/7340 [252:02<13:25, 27.7 steps/min]2025-08-11 19:38:21,487 - agent.ComputerAgent - INFO - LLM processing started with 22 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 22 messages\n", - "\u001b[92m19:38:21 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 19:38:22,165 - agent.ComputerAgent - INFO - LLM processing started with 18 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 18 messages\n", - "\u001b[92m19:38:22 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:38:22 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 95%|█████████████████████████████████████---| 6970/7340 [252:05<13:22, 27.6 steps/min]\u001b[92m19:38:23 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:38:24 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Loading checkpoint shards: 0%| | 0/4 [00:00/dev/null || echo 'Requesting sudo password if needed'\"})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': \"sudo -n true 2>/dev/null || echo 'Requesting sudo password if needed'\"})\n", - " 95%|██████████████████████████████████████--| 6973/7340 [252:29<13:17, 27.6 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/create_environment \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:38:48 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:38:49,132 - agent.ComputerAgent - INFO - Computer: click({'x': 178, 'y': 305})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 178, 'y': 305})\n", - " 95%|██████████████████████████████████████--| 6974/7340 [252:30<13:15, 27.6 steps/min]\u001b[92m19:38:49 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:38:49,800 - agent.ComputerAgent - INFO - Computer: click({'x': 932, 'y': 296})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 932, 'y': 296})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/fa081188-4859-4858-9d33-0f9675111182/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:38:50 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 95%|██████████████████████████████████████--| 6975/7340 [252:32<13:12, 27.6 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/create_environment \"HTTP/1.1 200 OK\"\n", - " 95%|██████████████████████████████████████--| 6976/7340 [252:33<13:10, 27.6 steps/min]\u001b[92m19:38:51 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:38:52,154 - agent.ComputerAgent - INFO - Computer: click({'x': 88, 'y': 313})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 88, 'y': 313})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/51c56274-d8ae-4edf-8ff1-b637cd2fff66/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:38:52 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/create_environment \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v1/environments/fa081188-4859-4858-9d33-0f9675111182/reset \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/cb64a220-43d8-4373-bd2a-e73bacb4a122/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - " 95%|██████████████████████████████████████--| 6976/7340 [252:35<13:10, 27.6 steps/min]\u001b[92m19:38:53 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/1a178f89-87e5-46d9-a114-22d5fcc5c630/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:38:54,176 - agent.ComputerAgent - INFO - LLM processing started with 12 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 12 messages\n", - "\u001b[92m19:38:54 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m19:38:54 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m19:38:54 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/8e75deb1-3c97-408b-8c7d-f4681b322141/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/68f513cf-ec21-4216-bab9-84c5bfcfce88/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/25f45afe-ee57-4629-9991-c515438accab/invoke \"HTTP/1.1 200 OK\"\n", - " 95%|██████████████████████████████████████--| 6977/7340 [252:36<13:08, 27.6 steps/min]2025-08-11 19:38:55,531 - agent.ComputerAgent - INFO - Agent: Taking a screenshot to see the current computer screen.\n", - "INFO:agent.ComputerAgent:Agent: Taking a screenshot to see the current computer screen.\n", - "2025-08-11 19:38:55,532 - agent.ComputerAgent - INFO - Computer: double_click({'x': 379, 'y': 105})\n", - "INFO:agent.ComputerAgent:Computer: double_click({'x': 379, 'y': 105})\n", - "2025-08-11 19:38:56,196 - agent.ComputerAgent - INFO - LLM processing started with 12 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 12 messages\n", - "\u001b[92m19:38:56 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 95%|██████████████████████████████████████--| 6977/7340 [252:37<13:08, 27.6 steps/min]2025-08-11 19:38:56,856 - agent.ComputerAgent - INFO - LLM processing started with 16 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 16 messages\n", - "\u001b[92m19:38:56 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m19:38:57 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:38:57,552 - agent.ComputerAgent - INFO - Computer: click({'x': 351, 'y': 75})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 351, 'y': 75})\n", - " 95%|██████████████████████████████████████--| 6978/7340 [252:39<13:06, 27.6 steps/min]2025-08-11 19:38:58,981 - agent.ComputerAgent - INFO - LLM processing started with 20 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 20 messages\n", - "\u001b[92m19:38:59 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/d71be89e-00e2-40e7-8b8d-38e36bc6d26c/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/fa081188-4859-4858-9d33-0f9675111182/invoke \"HTTP/1.1 200 OK\"\n", - " 95%|██████████████████████████████████████--| 6979/7340 [252:40<13:04, 27.6 steps/min]2025-08-11 19:38:59,669 - agent.ComputerAgent - INFO - LLM processing started with 1 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 1 messages\n", - "\u001b[92m19:38:59 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m19:38:59 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m19:38:59 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 95%|██████████████████████████████████████--| 6979/7340 [252:42<13:04, 27.6 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/40c5f987-3d81-47fe-8798-4e45d9755f93/invoke \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:39:02,357 - agent.ComputerAgent - INFO - LLM processing started with 6 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 6 messages\n", - "\u001b[92m19:39:02 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:39:02 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:39:03,722 - agent.ComputerAgent - INFO - Computer: wait({})\n", - "INFO:agent.ComputerAgent:Computer: wait({})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/ba3f752c-7268-49e8-819c-5b471e52bd54/invoke \"HTTP/1.1 200 OK\"\n", - " 95%|██████████████████████████████████████--| 6979/7340 [252:45<13:04, 27.6 steps/min]2025-08-11 19:39:04,403 - agent.ComputerAgent - INFO - Computer: scroll({'scroll_y': 640, 'scroll_x': 0, 'x': 993, 'y': 732})\n", - "INFO:agent.ComputerAgent:Computer: scroll({'scroll_y': 640, 'scroll_x': 0, 'x': 993, 'y': 732})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/6f3b006b-141d-439d-b6cb-eed7bd6483c3/invoke \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:39:05,068 - agent.ComputerAgent - INFO - LLM processing started with 16 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 16 messages\n", - "\u001b[92m19:39:05 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m19:39:05 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 95%|██████████████████████████████████████--| 6980/7340 [252:46<13:02, 27.6 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:39:06,128 - agent.ComputerAgent - INFO - Computer: click({'x': 316, 'y': 183})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 316, 'y': 183})\n", - "2025-08-11 19:39:06,806 - agent.ComputerAgent - INFO - LLM processing started with 24 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 24 messages\n", - "\u001b[92m19:39:06 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:39:08,174 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'ENTER'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'ENTER'})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - " 95%|██████████████████████████████████████--| 6981/7340 [252:50<13:00, 27.6 steps/min]\u001b[92m19:39:08 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:39:08 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:39:10,148 - agent.ComputerAgent - INFO - Agent: Taking a screenshot to see the current computer screen.\n", - "INFO:agent.ComputerAgent:Agent: Taking a screenshot to see the current computer screen.\n", - "2025-08-11 19:39:10,149 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'meta'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'meta'})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:39:10 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:39:11,496 - agent.ComputerAgent - INFO - Computer: click({'x': 569, 'y': 372})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 569, 'y': 372})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:39:11 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:39:12,817 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'ctrl+a'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'ctrl+a'})\n", - " 95%|██████████████████████████████████████--| 6983/7340 [252:54<12:55, 27.6 steps/min]2025-08-11 19:39:13,521 - agent.ComputerAgent - INFO - Computer: click({'x': 81, 'y': 148})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 81, 'y': 148})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:39:13 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m19:39:14 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/create_environment \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/25f45afe-ee57-4629-9991-c515438accab/invoke \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:39:14,827 - agent.ComputerAgent - INFO - Computer: drag({'path': [{'x': 210, 'y': 177}, {'x': 210, 'y': 457}]})\n", - "INFO:agent.ComputerAgent:Computer: drag({'path': [{'x': 210, 'y': 177}, {'x': 210, 'y': 457}]})\n", - "2025-08-11 19:39:15,498 - agent.ComputerAgent - INFO - LLM processing started with 14 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 14 messages\n", - "\u001b[92m19:39:15 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 19:39:16,177 - agent.ComputerAgent - INFO - LLM processing started with 14 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 14 messages\n", - " 95%|██████████████████████████████████████--| 6985/7340 [252:57<12:51, 27.6 steps/min]\u001b[92m19:39:16 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m19:39:16 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/create_environment \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:39:16 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 95%|██████████████████████████████████████--| 6987/7340 [252:58<12:46, 27.6 steps/min]\u001b[92m19:39:17 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:39:17,905 - agent.ComputerAgent - INFO - Computer: click({'x': 461, 'y': 321})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 461, 'y': 321})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/9882ec8e-4618-4be3-802e-bb5c58c9fbbc/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/b14fe395-5fa2-43f0-9d0b-23c42f3e9093/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:39:18 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 95%|██████████████████████████████████████--| 6988/7340 [253:01<12:44, 27.6 steps/min]\u001b[92m19:39:20 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:39:20,770 - agent.ComputerAgent - INFO - Computer: click({'x': 15, 'y': 430})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 15, 'y': 430})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/0180c5d2-a012-4261-b093-ed34f443f269/invoke \"HTTP/1.1 200 OK\"\n", - " 95%|██████████████████████████████████████--| 6988/7340 [253:02<12:44, 27.6 steps/min]2025-08-11 19:39:21,798 - agent.ComputerAgent - INFO - LLM processing started with 36 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 36 messages\n", - "\u001b[92m19:39:21 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/fa081188-4859-4858-9d33-0f9675111182/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/5e73167c-1836-4752-b7e8-57434e5d7875/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/68f513cf-ec21-4216-bab9-84c5bfcfce88/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/31367309-0055-409a-a992-edf729fb010c/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/c915dbd9-32bc-40a7-9c07-d437c737419f/invoke \"HTTP/1.1 200 OK\"\n", - " 95%|██████████████████████████████████████--| 6989/7340 [253:03<12:42, 27.6 steps/min]2025-08-11 19:39:22,498 - agent.ComputerAgent - INFO - LLM processing started with 6 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 6 messages\n", - "\u001b[92m19:39:22 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:39:23,147 - agent.ComputerAgent - INFO - LLM processing started with 20 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 20 messages\n", - "\u001b[92m19:39:23 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m19:39:23 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 95%|██████████████████████████████████████--| 6989/7340 [253:04<12:42, 27.6 steps/min]2025-08-11 19:39:23,865 - agent.ComputerAgent - INFO - Computer: drag({'path': [{'x': 209, 'y': 146}, {'x': 281, 'y': 396}]})\n", - "INFO:agent.ComputerAgent:Computer: drag({'path': [{'x': 209, 'y': 146}, {'x': 281, 'y': 396}]})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/0c64a3b4-e9b0-46c1-a580-cdcf62b74e44/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/7f112db6-0b60-4e6c-86f5-0d87dc91f371/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:39:24,516 - agent.ComputerAgent - INFO - LLM processing started with 20 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 20 messages\n", - "\u001b[92m19:39:24 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/1a178f89-87e5-46d9-a114-22d5fcc5c630/invoke \"HTTP/1.1 200 OK\"\n", - " 95%|██████████████████████████████████████--| 6989/7340 [253:06<12:42, 27.6 steps/min]2025-08-11 19:39:25,167 - agent.ComputerAgent - INFO - LLM processing started with 34 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 34 messages\n", - "\u001b[92m19:39:25 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:39:25 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:39:27,161 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'esc'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'esc'})\n", - "2025-08-11 19:39:27,846 - agent.ComputerAgent - INFO - LLM processing started with 32 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 32 messages\n", - "\u001b[92m19:39:27 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 19:39:28,870 - agent.ComputerAgent - INFO - LLM processing started with 42 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 42 messages\n", - "\u001b[92m19:39:28 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/1a178f89-87e5-46d9-a114-22d5fcc5c630/close \"HTTP/1.1 200 OK\"\n", - " 95%|██████████████████████████████████████--| 6990/7340 [253:10<12:40, 27.6 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m19:39:28 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:39:30,222 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'ctrl+a'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'ctrl+a'})\n", - "2025-08-11 19:39:30,916 - agent.ComputerAgent - INFO - Computer: click({'x': 91, 'y': 314})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 91, 'y': 314})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:39:32 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/40c5f987-3d81-47fe-8798-4e45d9755f93/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 95%|██████████████████████████████████████--| 6991/7340 [253:13<12:38, 27.6 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/acf3037a-4b6c-4ea8-b81c-ffc2e76132e1/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:39:32 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:39:33,548 - agent.ComputerAgent - INFO - LLM processing started with 16 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 16 messages\n", - "\u001b[92m19:39:33 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m19:39:33 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 95%|██████████████████████████████████████--| 6992/7340 [253:15<12:36, 27.6 steps/min]2025-08-11 19:39:34,246 - agent.ComputerAgent - INFO - LLM processing started with 22 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 22 messages\n", - "\u001b[92m19:39:34 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 19:39:34,914 - agent.ComputerAgent - INFO - Computer: click({'x': 341, 'y': 305})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 341, 'y': 305})\n", - "2025-08-11 19:39:35,577 - agent.ComputerAgent - INFO - LLM processing started with 8 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 8 messages\n", - "\u001b[92m19:39:35 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 95%|██████████████████████████████████████--| 6992/7340 [253:17<12:36, 27.6 steps/min]2025-08-11 19:39:36,257 - agent.ComputerAgent - INFO - LLM processing started with 26 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 26 messages\n", - "\u001b[92m19:39:36 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: GET https://orchestration.hud.so/hud-gym/api/v1/gyms/OSWorld-Ubuntu \"HTTP/1.1 200 OK\"\n", - "Loading checkpoint shards: 0%| | 0/4 [00:00/dev/null 2>&1; then\\n echo \\'User charles already exists\\'\\nelse\\n sudo useradd -M -s /usr/sbin/nologin -d / charles\\nfi\\n\\necho \\'charles:Ex@mpleP@55w0rd!\\' | sudo chpasswd\\n\\n# Give the user write access inside the jail via a subdirectory\\nsudo mkdir -p /home/test1/uploads\\nsudo chown charles:charles /home/test1/uploads\\n\\n# Configure OpenSSH to chroot the user and force SFTP only\\nsudo bash -lc \\'cat >>/etc/ssh/sshd_config <<\"EOF\"\\n\\nMatch User charles\\n ChrootDirectory /home/test1\\n ForceCommand internal-sftp\\n AllowTCPForwarding no\\n X11Forwarding no\\nEOF\\'\\n\\n# Restart SSH to apply changes\\nsudo systemctl restart ssh || sudo systemctl restart sshd\\n\\necho \\'Setup complete: user charles restricted to /home/test1 via SFTP.\\''})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': 'set -e\\nsudo mkdir -p /home/test1/uploads\\nsudo chown root:root /home/test1\\nsudo chmod 755 /home/test1\\n\\n# Create the user for SFTP-only access chrooted to /home/test1\\nif id -u charles >/dev/null 2>&1; then\\n echo \\'User charles already exists\\'\\nelse\\n sudo useradd -M -s /usr/sbin/nologin -d / charles\\nfi\\n\\necho \\'charles:Ex@mpleP@55w0rd!\\' | sudo chpasswd\\n\\n# Give the user write access inside the jail via a subdirectory\\nsudo mkdir -p /home/test1/uploads\\nsudo chown charles:charles /home/test1/uploads\\n\\n# Configure OpenSSH to chroot the user and force SFTP only\\nsudo bash -lc \\'cat >>/etc/ssh/sshd_config <<\"EOF\"\\n\\nMatch User charles\\n ChrootDirectory /home/test1\\n ForceCommand internal-sftp\\n AllowTCPForwarding no\\n X11Forwarding no\\nEOF\\'\\n\\n# Restart SSH to apply changes\\nsudo systemctl restart ssh || sudo systemctl restart sshd\\n\\necho \\'Setup complete: user charles restricted to /home/test1 via SFTP.\\''})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Loading checkpoint shards: 100%|██████████| 4/4 [00:05<00:00, 1.34s/it]\n", - "\u001b[92m19:41:20 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 96%|██████████████████████████████████████--| 7029/7340 [255:01<11:17, 27.6 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:41:21 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 96%|██████████████████████████████████████--| 7030/7340 [255:03<11:14, 27.6 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m19:41:22 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m19:41:22 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:41:22,776 - agent.ComputerAgent - INFO - Computer: click({'x': 187, 'y': 52})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 187, 'y': 52})\n", - " 96%|██████████████████████████████████████--| 7030/7340 [255:04<11:14, 27.6 steps/min]\u001b[92m19:41:22 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:41:23,473 - agent.ComputerAgent - INFO - Computer: click({'x': 540, 'y': 471})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 540, 'y': 471})\n", - "\u001b[92m19:41:23 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:41:24,156 - agent.ComputerAgent - INFO - Computer: click({'x': 865, 'y': 201})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 865, 'y': 201})\n", - "\u001b[92m19:41:24 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:41:24 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 96%|██████████████████████████████████████--| 7031/7340 [255:06<11:12, 27.6 steps/min]\u001b[92m19:41:24 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:41:25,511 - agent.ComputerAgent - INFO - Computer: click({'x': 91, 'y': 314, 'button': 'left'})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 91, 'y': 314, 'button': 'left'})\n", - "2025-08-11 19:41:26,190 - agent.ComputerAgent - INFO - Computer: scroll({'scroll_y': 640, 'scroll_x': 0, 'x': 990, 'y': 709})\n", - "INFO:agent.ComputerAgent:Computer: scroll({'scroll_y': 640, 'scroll_x': 0, 'x': 990, 'y': 709})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/25f45afe-ee57-4629-9991-c515438accab/invoke \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:41:26 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:41:26,861 - agent.ComputerAgent - INFO - Computer: click({'x': 13, 'y': 673})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 13, 'y': 673})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:41:28,168 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'F11'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'F11'})\n", - " 96%|██████████████████████████████████████--| 7033/7340 [255:09<11:08, 27.6 steps/min]2025-08-11 19:41:28,840 - agent.ComputerAgent - INFO - LLM processing started with 24 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 24 messages\n", - "\u001b[92m19:41:28 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/68f513cf-ec21-4216-bab9-84c5bfcfce88/invoke \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:41:28 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:41:29,543 - agent.ComputerAgent - INFO - Computer: click({'x': 461, 'y': 101})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 461, 'y': 101})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - " 96%|██████████████████████████████████████--| 7037/7340 [255:11<10:59, 27.6 steps/min]\u001b[92m19:41:30 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:41:30,890 - agent.ComputerAgent - INFO - LLM processing started with 28 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 28 messages\n", - "\u001b[92m19:41:30 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m19:41:30 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:41:31,564 - agent.ComputerAgent - INFO - Computer: click({'x': 510, 'y': 283})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 510, 'y': 283})\n", - " 96%|██████████████████████████████████████--| 7039/7340 [255:16<10:54, 27.6 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/31367309-0055-409a-a992-edf729fb010c/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/8e75deb1-3c97-408b-8c7d-f4681b322141/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/ba3f752c-7268-49e8-819c-5b471e52bd54/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/51c56274-d8ae-4edf-8ff1-b637cd2fff66/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:41:35,309 - agent.ComputerAgent - INFO - LLM processing started with 26 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 26 messages\n", - "\u001b[92m19:41:35 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/fa081188-4859-4858-9d33-0f9675111182/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/0c64a3b4-e9b0-46c1-a580-cdcf62b74e44/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:41:36,009 - agent.ComputerAgent - INFO - LLM processing started with 26 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 26 messages\n", - "\u001b[92m19:41:36 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/5e73167c-1836-4752-b7e8-57434e5d7875/invoke \"HTTP/1.1 200 OK\"\n", - " 96%|██████████████████████████████████████--| 7039/7340 [255:17<10:55, 27.6 steps/min]2025-08-11 19:41:36,709 - agent.ComputerAgent - INFO - LLM processing started with 18 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 18 messages\n", - "\u001b[92m19:41:36 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 19:41:37,392 - agent.ComputerAgent - INFO - LLM processing started with 24 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 24 messages\n", - "\u001b[92m19:41:37 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 19:41:38,081 - agent.ComputerAgent - INFO - LLM processing started with 22 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 22 messages\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:41:38 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m19:41:38 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:41:39 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/7f112db6-0b60-4e6c-86f5-0d87dc91f371/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/40c5f987-3d81-47fe-8798-4e45d9755f93/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - " 96%|██████████████████████████████████████--| 7039/7340 [255:21<10:55, 27.6 steps/min]\u001b[92m19:41:40 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m19:41:40 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m19:41:40 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:41:41,479 - agent.ComputerAgent - INFO - LLM processing started with 18 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 18 messages\n", - "\u001b[92m19:41:41 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m19:41:41 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:41:42,130 - agent.ComputerAgent - INFO - LLM processing started with 28 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 28 messages\n", - "\u001b[92m19:41:42 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m19:41:42 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 96%|██████████████████████████████████████--| 7039/7340 [255:23<10:55, 27.6 steps/min]\u001b[92m19:41:42 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:41:42,794 - agent.ComputerAgent - INFO - Computer: click({'x': 996, 'y': 732})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 996, 'y': 732})\n", - "2025-08-11 19:41:43,448 - agent.ComputerAgent - INFO - Computer: click({'x': 223, 'y': 35})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 223, 'y': 35})\n", - "\u001b[92m19:41:43 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m19:41:43 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 96%|██████████████████████████████████████--| 7039/7340 [255:25<10:55, 27.6 steps/min]2025-08-11 19:41:44,120 - agent.ComputerAgent - INFO - Computer: click({'x': 343, 'y': 195})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 343, 'y': 195})\n", - "2025-08-11 19:41:44,810 - agent.ComputerAgent - INFO - Computer: drag({'path': [{'x': 209, 'y': 146}, {'x': 407, 'y': 399}]})\n", - "INFO:agent.ComputerAgent:Computer: drag({'path': [{'x': 209, 'y': 146}, {'x': 407, 'y': 399}]})\n", - " 96%|██████████████████████████████████████--| 7041/7340 [255:26<10:50, 27.6 steps/min]2025-08-11 19:41:45,450 - agent.ComputerAgent - INFO - LLM processing started with 40 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 40 messages\n", - "\u001b[92m19:41:45 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 19:41:46,139 - agent.ComputerAgent - INFO - LLM processing started with 38 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 38 messages\n", - "\u001b[92m19:41:46 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 96%|██████████████████████████████████████--| 7043/7340 [255:30<10:46, 27.6 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:41:50,003 - agent.ComputerAgent - INFO - Computer: type({'text': 'chrome refresh 2023'})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': 'chrome refresh 2023'})\n", - " 96%|██████████████████████████████████████--| 7043/7340 [255:31<10:46, 27.6 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:41:50 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:41:51,944 - agent.ComputerAgent - INFO - Computer: type({'text': 'Yann LeCun Google Scholar'})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': 'Yann LeCun Google Scholar'})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/25f45afe-ee57-4629-9991-c515438accab/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/0180c5d2-a012-4261-b093-ed34f443f269/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/acf3037a-4b6c-4ea8-b81c-ffc2e76132e1/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/6f3b006b-141d-439d-b6cb-eed7bd6483c3/invoke \"HTTP/1.1 200 OK\"\n", - " 96%|██████████████████████████████████████--| 7044/7340 [255:33<10:44, 27.6 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:41:52,582 - agent.ComputerAgent - INFO - LLM processing started with 26 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 26 messages\n", - "\u001b[92m19:41:52 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:41:53 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m19:41:53 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - " 96%|██████████████████████████████████████--| 7045/7340 [255:35<10:42, 27.6 steps/min]\u001b[92m19:41:53 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:41:54,575 - agent.ComputerAgent - INFO - LLM processing started with 32 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 32 messages\n", - "\u001b[92m19:41:54 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 19:41:55,282 - agent.ComputerAgent - INFO - Computer: click({'x': 90, 'y': 314, 'button': 'left'})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 90, 'y': 314, 'button': 'left'})\n", - "\u001b[92m19:41:55 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v1/environments/26dc2412-0699-4a4e-a272-dc576348a5c8/reset \"HTTP/1.1 200 OK\"\n", - " 96%|██████████████████████████████████████--| 7045/7340 [255:37<10:42, 27.6 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:41:55,970 - agent.ComputerAgent - INFO - Computer: double_click({'x': 12, 'y': 524})\n", - "INFO:agent.ComputerAgent:Computer: double_click({'x': 12, 'y': 524})\n", - "\u001b[92m19:41:55 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:41:56,628 - agent.ComputerAgent - INFO - Computer: click({'x': 164, 'y': 427})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 164, 'y': 427})\n", - " 96%|██████████████████████████████████████--| 7046/7340 [255:38<10:40, 27.6 steps/min]2025-08-11 19:41:57,310 - agent.ComputerAgent - INFO - LLM processing started with 34 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 34 messages\n", - "\u001b[92m19:41:57 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:41:58 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 96%|██████████████████████████████████████--| 7048/7340 [255:40<10:35, 27.6 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m19:41:58 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:41:59,486 - agent.ComputerAgent - INFO - Computer: scroll({'scroll_y': 590, 'scroll_x': 0, 'x': 991, 'y': 420})\n", - "INFO:agent.ComputerAgent:Computer: scroll({'scroll_y': 590, 'scroll_x': 0, 'x': 991, 'y': 420})\n", - " 96%|██████████████████████████████████████--| 7048/7340 [255:41<10:35, 27.6 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/0180c5d2-a012-4261-b093-ed34f443f269/invoke \"HTTP/1.1 200 OK\"\n", - " 96%|██████████████████████████████████████--| 7049/7340 [255:42<10:33, 27.6 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/26dc2412-0699-4a4e-a272-dc576348a5c8/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:42:02,043 - agent.ComputerAgent - INFO - LLM processing started with 1 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 1 messages\n", - "\u001b[92m19:42:02 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/ba3f752c-7268-49e8-819c-5b471e52bd54/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/40c5f987-3d81-47fe-8798-4e45d9755f93/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/fa081188-4859-4858-9d33-0f9675111182/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/0180c5d2-a012-4261-b093-ed34f443f269/close \"HTTP/1.1 200 OK\"\n", - " 96%|██████████████████████████████████████--| 7049/7340 [255:43<10:33, 27.6 steps/min]2025-08-11 19:42:03,400 - agent.ComputerAgent - INFO - LLM processing started with 20 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 20 messages\n", - "\u001b[92m19:42:03 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 19:42:04,040 - agent.ComputerAgent - INFO - LLM processing started with 20 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 20 messages\n", - "\u001b[92m19:42:04 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:42:04 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m19:42:05 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/68f513cf-ec21-4216-bab9-84c5bfcfce88/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/51c56274-d8ae-4edf-8ff1-b637cd2fff66/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:42:06 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 96%|██████████████████████████████████████--| 7049/7340 [255:47<10:33, 27.6 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/5e73167c-1836-4752-b7e8-57434e5d7875/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:42:06,703 - agent.ComputerAgent - INFO - LLM processing started with 26 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 26 messages\n", - "\u001b[92m19:42:06 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "Loading checkpoint shards: 0%| | 0/4 [00:00Background Cover'})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': '>Background Cover'})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/7f112db6-0b60-4e6c-86f5-0d87dc91f371/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/create_environment \"HTTP/1.1 200 OK\"\n", - " 96%|██████████████████████████████████████--| 7072/7340 [256:50<9:43, 27.5 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/9882ec8e-4618-4be3-802e-bb5c58c9fbbc/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:43:08,956 - agent.ComputerAgent - INFO - LLM processing started with 1 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 1 messages\n", - "\u001b[92m19:43:08 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:43:10,282 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'ENTER'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'ENTER'})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:43:10 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:43:12,251 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'ENTER'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'ENTER'})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m19:43:12 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/ba3f752c-7268-49e8-819c-5b471e52bd54/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 96%|██████████████████████████████████████--| 7074/7340 [256:54<9:39, 27.5 steps/min]2025-08-11 19:43:13,540 - agent.ComputerAgent - INFO - LLM processing started with 1 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 1 messages\n", - "\u001b[92m19:43:13 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m19:43:13 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:43:14,917 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'enter'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'enter'})\n", - "2025-08-11 19:43:15,565 - agent.ComputerAgent - INFO - Agent: Taking a screenshot to see the current computer screen.\n", - "INFO:agent.ComputerAgent:Agent: Taking a screenshot to see the current computer screen.\n", - "2025-08-11 19:43:15,566 - agent.ComputerAgent - INFO - Computer: click({'x': 75, 'y': 166})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 75, 'y': 166})\n", - "\u001b[92m19:43:15 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 96%|██████████████████████████████████████--| 7076/7340 [256:57<9:35, 27.5 steps/min]2025-08-11 19:43:16,239 - agent.ComputerAgent - INFO - Computer: click({'x': 985, 'y': 759})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 985, 'y': 759})\n", - "2025-08-11 19:43:16,881 - agent.ComputerAgent - INFO - LLM processing started with 42 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 42 messages\n", - "\u001b[92m19:43:16 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 96%|██████████████████████████████████████--| 7078/7340 [256:58<9:30, 27.5 steps/min]2025-08-11 19:43:17,557 - agent.ComputerAgent - INFO - LLM processing started with 32 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 32 messages\n", - "\u001b[92m19:43:17 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:43:18 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/765aa707-cf44-4dd0-8933-2c2b94870afd/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:43:18,888 - agent.ComputerAgent - INFO - LLM processing started with 38 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 38 messages\n", - "\u001b[92m19:43:18 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 96%|██████████████████████████████████████--| 7079/7340 [257:00<9:28, 27.5 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:43:20,265 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'ctrl++'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'ctrl++'})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:43:20 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 96%|██████████████████████████████████████--| 7079/7340 [257:02<9:28, 27.5 steps/min]\u001b[92m19:43:20 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:43:21,621 - agent.ComputerAgent - INFO - Computer: scroll({'scroll_y': 588, 'scroll_x': 0, 'x': 991, 'y': 433})\n", - "INFO:agent.ComputerAgent:Computer: scroll({'scroll_y': 588, 'scroll_x': 0, 'x': 991, 'y': 433})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:43:22 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/51c56274-d8ae-4edf-8ff1-b637cd2fff66/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:43:23,319 - agent.ComputerAgent - INFO - LLM processing started with 26 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 26 messages\n", - "\u001b[92m19:43:23 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/25f45afe-ee57-4629-9991-c515438accab/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/d71be89e-00e2-40e7-8b8d-38e36bc6d26c/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m19:43:24 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/26dc2412-0699-4a4e-a272-dc576348a5c8/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/68f513cf-ec21-4216-bab9-84c5bfcfce88/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/0c64a3b4-e9b0-46c1-a580-cdcf62b74e44/invoke \"HTTP/1.1 200 OK\"\n", - " 96%|██████████████████████████████████████--| 7079/7340 [257:05<9:28, 27.5 steps/min]\u001b[92m19:43:24 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:43:24,668 - agent.ComputerAgent - INFO - LLM processing started with 32 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 32 messages\n", - "\u001b[92m19:43:24 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 19:43:25,366 - agent.ComputerAgent - INFO - Computer: click({'x': 341, 'y': 75})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 341, 'y': 75})\n", - "\u001b[92m19:43:25 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m19:43:25 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:43:26,376 - agent.ComputerAgent - INFO - Agent: Taking a screenshot to see the current computer screen.\n", - "INFO:agent.ComputerAgent:Agent: Taking a screenshot to see the current computer screen.\n", - "2025-08-11 19:43:26,377 - agent.ComputerAgent - INFO - Computer: click({'x': 48, 'y': 52})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 48, 'y': 52})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:43:27,706 - agent.ComputerAgent - INFO - Computer: click({'x': 213, 'y': 183})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 213, 'y': 183})\n", - " 96%|██████████████████████████████████████--| 7080/7340 [257:09<9:26, 27.5 steps/min]2025-08-11 19:43:28,338 - agent.ComputerAgent - INFO - LLM processing started with 30 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 30 messages\n", - "\u001b[92m19:43:28 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 19:43:29,016 - agent.ComputerAgent - INFO - Computer: click({'x': 666, 'y': 279})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 666, 'y': 279})\n", - " 96%|██████████████████████████████████████--| 7083/7340 [257:10<9:19, 27.5 steps/min]2025-08-11 19:43:29,687 - agent.ComputerAgent - INFO - LLM processing started with 6 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 6 messages\n", - "\u001b[92m19:43:29 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 19:43:30,358 - agent.ComputerAgent - INFO - LLM processing started with 10 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 10 messages\n", - "\u001b[92m19:43:30 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/cb64a220-43d8-4373-bd2a-e73bacb4a122/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 97%|██████████████████████████████████████--| 7084/7340 [257:12<9:17, 27.5 steps/min]2025-08-11 19:43:31,038 - agent.ComputerAgent - INFO - LLM processing started with 8 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 8 messages\n", - "\u001b[92m19:43:31 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:43:31 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v1/environments/765aa707-cf44-4dd0-8933-2c2b94870afd/reset \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 97%|██████████████████████████████████████--| 7084/7340 [257:13<9:17, 27.5 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:43:32,348 - agent.ComputerAgent - INFO - LLM processing started with 34 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 34 messages\n", - "\u001b[92m19:43:32 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m19:43:32 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:43:33,008 - agent.ComputerAgent - INFO - Agent: Taking a screenshot to see the current computer screen.\n", - "INFO:agent.ComputerAgent:Agent: Taking a screenshot to see the current computer screen.\n", - "2025-08-11 19:43:33,009 - agent.ComputerAgent - INFO - Computer: click({'x': 79, 'y': 157})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 79, 'y': 157})\n", - " 97%|██████████████████████████████████████--| 7084/7340 [257:14<9:17, 27.5 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:43:34 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/0c64a3b4-e9b0-46c1-a580-cdcf62b74e44/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 97%|██████████████████████████████████████--| 7085/7340 [257:15<9:15, 27.5 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m19:43:34 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:43:35,405 - agent.ComputerAgent - INFO - Computer: click({'x': 679, 'y': 563})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 679, 'y': 563})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:43:36,736 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'ctrl++'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'ctrl++'})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/6f3b006b-141d-439d-b6cb-eed7bd6483c3/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:43:37 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/9882ec8e-4618-4be3-802e-bb5c58c9fbbc/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/fa081188-4859-4858-9d33-0f9675111182/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/0c64a3b4-e9b0-46c1-a580-cdcf62b74e44/close \"HTTP/1.1 200 OK\"\n", - " 97%|██████████████████████████████████████--| 7086/7340 [257:19<9:13, 27.5 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/765aa707-cf44-4dd0-8933-2c2b94870afd/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/5e73167c-1836-4752-b7e8-57434e5d7875/invoke \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:43:38,000 - agent.ComputerAgent - INFO - LLM processing started with 28 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 28 messages\n", - "\u001b[92m19:43:38 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/31367309-0055-409a-a992-edf729fb010c/invoke \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:43:38 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/b14fe395-5fa2-43f0-9d0b-23c42f3e9093/invoke \"HTTP/1.1 200 OK\"\n", - " 97%|██████████████████████████████████████--| 7087/7340 [257:20<9:11, 27.5 steps/min]2025-08-11 19:43:39,348 - agent.ComputerAgent - INFO - Computer: click({'button': 'left', 'x': 308, 'y': 153})\n", - "INFO:agent.ComputerAgent:Computer: click({'button': 'left', 'x': 308, 'y': 153})\n", - "2025-08-11 19:43:40,018 - agent.ComputerAgent - INFO - LLM processing started with 26 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 26 messages\n", - "\u001b[92m19:43:40 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 19:43:40,687 - agent.ComputerAgent - INFO - LLM processing started with 6 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 6 messages\n", - "\u001b[92m19:43:40 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 19:43:41,369 - agent.ComputerAgent - INFO - LLM processing started with 36 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 36 messages\n", - "\u001b[92m19:43:41 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - " 97%|██████████████████████████████████████--| 7087/7340 [257:23<9:11, 27.5 steps/min]\u001b[92m19:43:42 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:43:42 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:43:43,783 - agent.ComputerAgent - INFO - LLM processing started with 34 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 34 messages\n", - "\u001b[92m19:43:43 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "Loading checkpoint shards: 0%| | 0/4 [00:00>/etc/ssh/sshd_config <<\"EOF\"\\n\\nMatch User charles\\n ChrootDirectory /home/test1\\n ForceCommand internal-sftp\\n AllowTCPForwarding no\\n X11Forwarding no\\nEOF\\'\\n\\n# Apply changes\\nsudo systemctl restart ssh || sudo systemctl restart sshd\\n\\n# Show resulting setup\\nid charles\\nls -ld /home/test1 /home/test1/uploads'})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': 'set -e\\nsudo mkdir -p /home/test1\\nsudo chown root:root /home/test1\\nsudo chmod 755 /home/test1\\n\\n# Create user with no shell access\\nsudo useradd -M -s /usr/sbin/nologin -d / charles || true\\n\\necho \\'charles:Ex@mpleP@55w0rd!\\' | sudo chpasswd\\n\\n# Writable directory inside the jail\\nsudo mkdir -p /home/test1/uploads\\nsudo chown charles:charles /home/test1/uploads\\n\\n# Configure SSH to restrict user to SFTP chroot at /home/test1\\nsudo bash -lc \\'cat >>/etc/ssh/sshd_config <<\"EOF\"\\n\\nMatch User charles\\n ChrootDirectory /home/test1\\n ForceCommand internal-sftp\\n AllowTCPForwarding no\\n X11Forwarding no\\nEOF\\'\\n\\n# Apply changes\\nsudo systemctl restart ssh || sudo systemctl restart sshd\\n\\n# Show resulting setup\\nid charles\\nls -ld /home/test1 /home/test1/uploads'})\n", - " 97%|██████████████████████████████████████--| 7110/7340 [257:49<8:20, 27.6 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:44:09 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/9882ec8e-4618-4be3-802e-bb5c58c9fbbc/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/fa081188-4859-4858-9d33-0f9675111182/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/40c5f987-3d81-47fe-8798-4e45d9755f93/invoke \"HTTP/1.1 200 OK\"\n", - " 97%|██████████████████████████████████████--| 7112/7340 [257:50<8:15, 27.6 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/acf3037a-4b6c-4ea8-b81c-ffc2e76132e1/invoke \"HTTP/1.1 200 OK\"\n", - "Loading checkpoint shards: 0%| | 0/4 [00:00 Settings > General > scroll to the bottom > Config Editor…\n", - "- In the search box, type: applyIncomingFilters\n", - "- If mail.server.default.applyIncomingFilters is false or missing, set it to true.\n", - "- If you prefer per-account control, also set mail.server.serverN.applyIncomingFilters = true for the specific account (N is the server number for that account; you can set both default and per-account if you like).\n", - "\n", - "2) Ensure Thunderbird actually checks those subfolders for new mail\n", - "- For each subfolder you want filters to run on: right‑click the folder > Properties > check “When getting new messages, always check this folder” > OK.\n", - "- Alternatively (account-wide): Account Settings > Server Settings (for the IMAP account) > enable “Check for new messages at startup” and “Check for new messages every X minutes.” In Advanced IMAP settings (or via Config Editor), ensure mail.server.serverN.check_all_folders_for_new = true to have Thunderbird poll all folders.\n", - "\n", - "3) Make sure your filters are set to run on incoming mail\n", - "- Tools (≡ menu) > Message Filters.\n", - "- Choose the account, edit each filter you want, and under “Apply filter when,” tick the “Getting New Mail” options (before/after junk classification as appropriate). Save.\n", - "\n", - "4) Restart Thunderbird\n", - "- Close and reopen Thunderbird so the new prefs apply cleanly.\n", - "\n", - "5) Test\n", - "- Deliver a message directly to a subfolder (e.g., by a server-side rule or by sending to an address that your server routes into that folder). When Thunderbird detects new mail in that subfolder, your filters should run automatically.\n", "\n", - "Notes and limitations\n", - "- IMAP required: This works when new mail arrives directly into that subfolder on the server (IMAP). For POP accounts and Local Folders, Thunderbird only “receives” mail into Inbox; filters won’t auto-trigger simply because a message...\n", - "INFO:agent.ComputerAgent:Agent: Yes—you can make Thunderbird run filters automatically on subfolders (not just Inbox). Do the following once per account you want this behavior for.\n", - "\n", - "1) Turn on the hidden preference to run filters on non-Inbox folders\n", - "- Thunderbird > Settings > General > scroll to the bottom > Config Editor…\n", - "- In the search box, type: applyIncomingFilters\n", - "- If mail.server.default.applyIncomingFilters is false or missing, set it to true.\n", - "- If you prefer per-account control, also set mail.server.serverN.applyIncomingFilters = true for the specific account (N is the server number for that account; you can set both default and per-account if you like).\n", - "\n", - "2) Ensure Thunderbird actually checks those subfolders for new mail\n", - "- For each subfolder you want filters to run on: right‑click the folder > Properties > check “When getting new messages, always check this folder” > OK.\n", - "- Alternatively (account-wide): Account Settings > Server Settings (for the IMAP account) > enable “Check for new messages at startup” and “Check for new messages every X minutes.” In Advanced IMAP settings (or via Config Editor), ensure mail.server.serverN.check_all_folders_for_new = true to have Thunderbird poll all folders.\n", - "\n", - "3) Make sure your filters are set to run on incoming mail\n", - "- Tools (≡ menu) > Message Filters.\n", - "- Choose the account, edit each filter you want, and under “Apply filter when,” tick the “Getting New Mail” options (before/after junk classification as appropriate). Save.\n", - "\n", - "4) Restart Thunderbird\n", - "- Close and reopen Thunderbird so the new prefs apply cleanly.\n", - "\n", - "5) Test\n", - "- Deliver a message directly to a subfolder (e.g., by a server-side rule or by sending to an address that your server routes into that folder). When Thunderbird detects new mail in that subfolder, your filters should run automatically.\n", - "\n", - "Notes and limitations\n", - "- IMAP required: This works when new mail arrives directly into that subfolder on the server (IMAP). For POP accounts and Local Folders, Thunderbird only “receives” mail into Inbox; filters won’t auto-trigger simply because a message...\n", - "2025-08-11 19:47:20,255 - agent.ComputerAgent - INFO - Total usage:\n", - " - completion_tokens: 3021\n", - " - prompt_tokens: 3980\n", - " - total_tokens: 7001\n", - " - completion_tokens_details:\n", - " - accepted_prediction_tokens: 0\n", - " - audio_tokens: 0\n", - " - reasoning_tokens: 2496\n", - " - rejected_prediction_tokens: 0\n", - " - prompt_tokens_details:\n", - " - audio_tokens: 0\n", - " - cached_tokens: 1920\n", - " - response_cost: $0.0330\n", - "INFO:agent.ComputerAgent:Total usage:\n", - " - completion_tokens: 3021\n", - " - prompt_tokens: 3980\n", - " - total_tokens: 7001\n", - " - completion_tokens_details:\n", - " - accepted_prediction_tokens: 0\n", - " - audio_tokens: 0\n", - " - reasoning_tokens: 2496\n", - " - rejected_prediction_tokens: 0\n", - " - prompt_tokens_details:\n", - " - audio_tokens: 0\n", - " - cached_tokens: 1920\n", - " - response_cost: $0.0330\n", - " 98%|███████████████████████████████████████-| 7212/7340 [261:01<4:37, 27.6 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:47:20 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m19:47:20 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:47:21,621 - agent.ComputerAgent - INFO - Computer: type({'text': ' chrome refresh 2023'})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': ' chrome refresh 2023'})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/5e73167c-1836-4752-b7e8-57434e5d7875/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:47:22,286 - agent.ComputerAgent - INFO - Computer: click({'x': 239, 'y': 178})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 239, 'y': 178})\n", - "2025-08-11 19:47:22,928 - agent.ComputerAgent - INFO - Computer: click({'x': 980, 'y': 60})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 980, 'y': 60})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/40c5f987-3d81-47fe-8798-4e45d9755f93/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/b14fe395-5fa2-43f0-9d0b-23c42f3e9093/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/cb64a220-43d8-4373-bd2a-e73bacb4a122/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/25f45afe-ee57-4629-9991-c515438accab/invoke \"HTTP/1.1 200 OK\"\n", - " 98%|███████████████████████████████████████-| 7213/7340 [261:04<4:35, 27.6 steps/min]2025-08-11 19:47:23,594 - agent.ComputerAgent - INFO - LLM processing started with 42 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 42 messages\n", - "\u001b[92m19:47:23 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 19:47:24,237 - agent.ComputerAgent - INFO - LLM processing started with 16 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 16 messages\n", - "\u001b[92m19:47:24 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 98%|███████████████████████████████████████-| 7216/7340 [261:06<4:29, 27.6 steps/min]2025-08-11 19:47:24,914 - agent.ComputerAgent - INFO - LLM processing started with 22 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 22 messages\n", - "\u001b[92m19:47:24 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:47:26,264 - agent.ComputerAgent - INFO - Agent: The Impress slide was exported to PNG using the default settings and saved as “res.png” on the Desktop. Task completed.\n", - "INFO:agent.ComputerAgent:Agent: The Impress slide was exported to PNG using the default settings and saved as “res.png” on the Desktop. Task completed.\n", - "2025-08-11 19:47:26,906 - agent.ComputerAgent - INFO - Total usage:\n", - " - completion_tokens: 162\n", - " - prompt_tokens: 8778\n", - " - total_tokens: 8940\n", - " - completion_tokens_details:\n", - " - accepted_prediction_tokens: 0\n", - " - audio_tokens: 0\n", - " - reasoning_tokens: 128\n", - " - rejected_prediction_tokens: 0\n", - " - prompt_tokens_details:\n", - " - audio_tokens: 0\n", - " - cached_tokens: 6656\n", - " - response_cost: $0.0051\n", - "INFO:agent.ComputerAgent:Total usage:\n", - " - completion_tokens: 162\n", - " - prompt_tokens: 8778\n", - " - total_tokens: 8940\n", - " - completion_tokens_details:\n", - " - accepted_prediction_tokens: 0\n", - " - audio_tokens: 0\n", - " - reasoning_tokens: 128\n", - " - rejected_prediction_tokens: 0\n", - " - prompt_tokens_details:\n", - " - audio_tokens: 0\n", - " - cached_tokens: 6656\n", - " - response_cost: $0.0051\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/cd53d966-2507-485c-bbe9-ee55dbbeddd0/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/5e73167c-1836-4752-b7e8-57434e5d7875/close \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/25f45afe-ee57-4629-9991-c515438accab/invoke \"HTTP/1.1 200 OK\"\n", - " 98%|███████████████████████████████████████-| 7217/7340 [261:09<4:27, 27.6 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/25f45afe-ee57-4629-9991-c515438accab/close \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/9882ec8e-4618-4be3-802e-bb5c58c9fbbc/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/6f3b006b-141d-439d-b6cb-eed7bd6483c3/invoke \"HTTP/1.1 200 OK\"\n", - " 98%|███████████████████████████████████████-| 7217/7340 [261:10<4:27, 27.6 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/cd53d966-2507-485c-bbe9-ee55dbbeddd0/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/fa081188-4859-4858-9d33-0f9675111182/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:47:30,124 - agent.ComputerAgent - INFO - LLM processing started with 40 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 40 messages\n", - "\u001b[92m19:47:30 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/9053f5ae-149f-4a73-a89e-977f3e750435/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:47:30 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/31367309-0055-409a-a992-edf729fb010c/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 99%|███████████████████████████████████████-| 7233/7340 [261:12<3:51, 27.7 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/cd53d966-2507-485c-bbe9-ee55dbbeddd0/close \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:47:31,452 - agent.ComputerAgent - INFO - LLM processing started with 10 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 10 messages\n", - "\u001b[92m19:47:31 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 19:47:32,084 - agent.ComputerAgent - INFO - LLM processing started with 42 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 42 messages\n", - "\u001b[92m19:47:32 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m19:47:32 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 99%|███████████████████████████████████████-| 7233/7340 [261:13<3:51, 27.7 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/b4eee866-c191-4acf-b232-9b18a3c888ef/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:47:32,764 - agent.ComputerAgent - INFO - Computer: click({'x': 76, 'y': 53})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 76, 'y': 53})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/6f3b006b-141d-439d-b6cb-eed7bd6483c3/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:47:33,390 - agent.ComputerAgent - INFO - LLM processing started with 12 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 12 messages\n", - "\u001b[92m19:47:33 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/9882ec8e-4618-4be3-802e-bb5c58c9fbbc/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 99%|███████████████████████████████████████-| 7235/7340 [261:15<3:47, 27.7 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/6f3b006b-141d-439d-b6cb-eed7bd6483c3/close \"HTTP/1.1 200 OK\"\n", - " 99%|███████████████████████████████████████-| 7245/7340 [261:16<3:25, 27.7 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/9882ec8e-4618-4be3-802e-bb5c58c9fbbc/close \"HTTP/1.1 200 OK\"\n", - " 99%|███████████████████████████████████████-| 7245/7340 [261:17<3:25, 27.7 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:47:36 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:47:36 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 99%|███████████████████████████████████████-| 7245/7340 [261:18<3:25, 27.7 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m19:47:36 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:47:37,446 - agent.ComputerAgent - INFO - Computer: click({'x': 901, 'y': 579})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 901, 'y': 579})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:47:38 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m19:47:38 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 99%|███████████████████████████████████████-| 7245/7340 [261:20<3:25, 27.7 steps/min]\u001b[92m19:47:38 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:47:39,450 - agent.ComputerAgent - INFO - Computer: click({'x': 1011, 'y': 194})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 1011, 'y': 194})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m19:47:40 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/26dc2412-0699-4a4e-a272-dc576348a5c8/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 99%|███████████████████████████████████████-| 7246/7340 [261:21<3:23, 27.7 steps/min]\u001b[92m19:47:40 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:47:40,815 - agent.ComputerAgent - INFO - Computer: click({'x': 243, 'y': 52})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 243, 'y': 52})\n", - "\u001b[92m19:47:40 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:47:41,487 - agent.ComputerAgent - INFO - Computer: click({'x': 259, 'y': 178})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 259, 'y': 178})\n", - "\u001b[92m19:47:41 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 99%|███████████████████████████████████████-| 7247/7340 [261:23<3:21, 27.7 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:47:42,158 - agent.ComputerAgent - INFO - Computer: click({'x': 893, 'y': 296})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 893, 'y': 296})\n", - "2025-08-11 19:47:42,787 - agent.ComputerAgent - INFO - LLM processing started with 26 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 26 messages\n", - "\u001b[92m19:47:42 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 99%|███████████████████████████████████████-| 7250/7340 [261:28<3:14, 27.7 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:47:47 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/8e75deb1-3c97-408b-8c7d-f4681b322141/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/9053f5ae-149f-4a73-a89e-977f3e750435/invoke \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:47:48,590 - agent.ComputerAgent - INFO - LLM processing started with 12 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 12 messages\n", - "\u001b[92m19:47:48 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m19:47:48 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/ba3f752c-7268-49e8-819c-5b471e52bd54/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/d71be89e-00e2-40e7-8b8d-38e36bc6d26c/invoke \"HTTP/1.1 200 OK\"\n", - " 99%|███████████████████████████████████████-| 7250/7340 [261:30<3:14, 27.7 steps/min]\u001b[92m19:47:48 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:47:49,265 - agent.ComputerAgent - INFO - LLM processing started with 38 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 38 messages\n", - "\u001b[92m19:47:49 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m19:47:49 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/b4eee866-c191-4acf-b232-9b18a3c888ef/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 99%|███████████████████████████████████████-| 7250/7340 [261:31<3:14, 27.7 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:47:50,594 - agent.ComputerAgent - INFO - LLM processing started with 14 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 14 messages\n", - "\u001b[92m19:47:50 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m19:47:50 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:47:51,304 - agent.ComputerAgent - INFO - LLM processing started with 20 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 20 messages\n", - "\u001b[92m19:47:51 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 19:47:51,990 - agent.ComputerAgent - INFO - Computer: drag({'path': [{'x': 275, 'y': 181}, {'x': 79, 'y': 182}]})\n", - "INFO:agent.ComputerAgent:Computer: drag({'path': [{'x': 275, 'y': 181}, {'x': 79, 'y': 182}]})\n", - " 99%|███████████████████████████████████████-| 7250/7340 [261:33<3:14, 27.7 steps/min]\u001b[92m19:47:51 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:47:52,662 - agent.ComputerAgent - INFO - Computer: click({'x': 252, 'y': 230})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 252, 'y': 230})\n", - " 99%|███████████████████████████████████████-| 7252/7340 [261:35<3:10, 27.7 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:47:54 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:47:56,131 - agent.ComputerAgent - INFO - Computer: type({'text': '=A2+B2'})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': '=A2+B2'})\n", - " 99%|███████████████████████████████████████-| 7252/7340 [261:37<3:10, 27.7 steps/min]\u001b[92m19:47:56 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:47:56,793 - agent.ComputerAgent - INFO - Computer: click({'x': 563, 'y': 101})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 563, 'y': 101})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:47:58,146 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'alt+left'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'alt+left'})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/cb64a220-43d8-4373-bd2a-e73bacb4a122/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:47:58 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/b14fe395-5fa2-43f0-9d0b-23c42f3e9093/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 99%|███████████████████████████████████████-| 7253/7340 [261:40<3:08, 27.7 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/ba3f752c-7268-49e8-819c-5b471e52bd54/invoke \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:47:59,467 - agent.ComputerAgent - INFO - LLM processing started with 18 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 18 messages\n", - "\u001b[92m19:47:59 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 19:48:00,135 - agent.ComputerAgent - INFO - LLM processing started with 24 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 24 messages\n", - "\u001b[92m19:48:00 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m19:48:00 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 99%|███████████████████████████████████████-| 7254/7340 [261:41<3:06, 27.7 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:48:00,837 - agent.ComputerAgent - INFO - Computer: click({'x': 111, 'y': 52})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 111, 'y': 52})\n", - " 99%|███████████████████████████████████████-| 7254/7340 [261:42<3:06, 27.7 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/ba3f752c-7268-49e8-819c-5b471e52bd54/close \"HTTP/1.1 200 OK\"\n", - " 99%|███████████████████████████████████████-| 7255/7340 [261:44<3:04, 27.7 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:48:03 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - " 99%|███████████████████████████████████████-| 7255/7340 [261:45<3:04, 27.7 steps/min]\u001b[92m19:48:04 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:48:04,749 - agent.ComputerAgent - INFO - Computer: click({'x': 19, 'y': 45})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 19, 'y': 45})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:48:05 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 99%|███████████████████████████████████████-| 7255/7340 [261:47<3:04, 27.7 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/fa081188-4859-4858-9d33-0f9675111182/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/9053f5ae-149f-4a73-a89e-977f3e750435/invoke \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:48:06,105 - agent.ComputerAgent - INFO - LLM processing started with 42 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 42 messages\n", - "\u001b[92m19:48:06 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 19:48:06,746 - agent.ComputerAgent - INFO - LLM processing started with 14 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 14 messages\n", - "\u001b[92m19:48:06 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 99%|███████████████████████████████████████-| 7256/7340 [261:48<3:01, 27.7 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m19:48:06 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:48:07,447 - agent.ComputerAgent - INFO - Computer: click({'x': 273, 'y': 90})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 273, 'y': 90})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/26dc2412-0699-4a4e-a272-dc576348a5c8/invoke \"HTTP/1.1 200 OK\"\n", - " 99%|███████████████████████████████████████-| 7256/7340 [261:49<3:01, 27.7 steps/min]2025-08-11 19:48:08,626 - agent.ComputerAgent - INFO - LLM processing started with 28 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 28 messages\n", - "\u001b[92m19:48:08 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 99%|███████████████████████████████████████-| 7257/7340 [261:51<2:59, 27.7 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:48:10 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/b4eee866-c191-4acf-b232-9b18a3c888ef/invoke \"HTTP/1.1 200 OK\"\n", - " 99%|███████████████████████████████████████-| 7257/7340 [261:52<2:59, 27.7 steps/min]2025-08-11 19:48:11,495 - agent.ComputerAgent - INFO - LLM processing started with 16 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 16 messages\n", - "\u001b[92m19:48:11 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m19:48:11 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m19:48:11 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - " 99%|███████████████████████████████████████-| 7257/7340 [261:53<2:59, 27.7 steps/min]\u001b[92m19:48:12 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:48:12,687 - agent.ComputerAgent - INFO - Computer: drag({'path': [{'x': 211, 'y': 178}, {'x': 211, 'y': 473}]})\n", - "INFO:agent.ComputerAgent:Computer: drag({'path': [{'x': 211, 'y': 178}, {'x': 211, 'y': 473}]})\n", - " 99%|███████████████████████████████████████-| 7257/7340 [261:54<2:59, 27.7 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/d71be89e-00e2-40e7-8b8d-38e36bc6d26c/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:48:14,357 - agent.ComputerAgent - INFO - LLM processing started with 22 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 22 messages\n", - "\u001b[92m19:48:14 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 99%|███████████████████████████████████████-| 7258/7340 [261:56<2:57, 27.7 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/40c5f987-3d81-47fe-8798-4e45d9755f93/invoke \"HTTP/1.1 200 OK\"\n", - " 99%|███████████████████████████████████████-| 7263/7340 [261:57<2:46, 27.7 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:48:17,188 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'enter'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'enter'})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:48:17 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/40c5f987-3d81-47fe-8798-4e45d9755f93/close \"HTTP/1.1 200 OK\"\n", - " 99%|███████████████████████████████████████-| 7263/7340 [261:59<2:46, 27.7 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/31367309-0055-409a-a992-edf729fb010c/invoke \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:48:18 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:48:19,057 - agent.ComputerAgent - INFO - Computer: click({'x': 257, 'y': 152})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 257, 'y': 152})\n", - " 99%|███████████████████████████████████████-| 7265/7340 [262:03<2:42, 27.7 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/31367309-0055-409a-a992-edf729fb010c/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/9053f5ae-149f-4a73-a89e-977f3e750435/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:48:23,287 - agent.ComputerAgent - INFO - LLM processing started with 16 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 16 messages\n", - "\u001b[92m19:48:23 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m19:48:23 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/31367309-0055-409a-a992-edf729fb010c/close \"HTTP/1.1 200 OK\"\n", - " 99%|███████████████████████████████████████-| 7265/7340 [262:05<2:42, 27.7 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m19:48:24 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/b14fe395-5fa2-43f0-9d0b-23c42f3e9093/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:48:25 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/68f513cf-ec21-4216-bab9-84c5bfcfce88/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m19:48:25 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 99%|███████████████████████████████████████-| 7265/7340 [262:07<2:42, 27.7 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:48:25,935 - agent.ComputerAgent - INFO - Computer: click({'x': 95, 'y': 74})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 95, 'y': 74})\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:48:26,606 - agent.ComputerAgent - INFO - LLM processing started with 20 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 20 messages\n", - "\u001b[92m19:48:26 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 99%|███████████████████████████████████████-| 7266/7340 [262:08<2:40, 27.7 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m19:48:26 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:48:27,282 - agent.ComputerAgent - INFO - Computer: click({'x': 986, 'y': 133})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 986, 'y': 133})\n", - "\u001b[92m19:48:27 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:48:27,901 - agent.ComputerAgent - INFO - Computer: click({'x': 528, 'y': 50})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 528, 'y': 50})\n", - " 99%|███████████████████████████████████████-| 7267/7340 [262:09<2:38, 27.7 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:48:29 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/68f513cf-ec21-4216-bab9-84c5bfcfce88/close \"HTTP/1.1 200 OK\"\n", - " 99%|███████████████████████████████████████-| 7269/7340 [262:10<2:33, 27.7 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m19:48:29 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:48:30,287 - agent.ComputerAgent - INFO - Computer: click({'x': 105, 'y': 230})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 105, 'y': 230})\n", - " 99%|███████████████████████████████████████-| 7270/7340 [262:14<2:31, 27.7 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/8e75deb1-3c97-408b-8c7d-f4681b322141/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:48:33,466 - agent.ComputerAgent - INFO - LLM processing started with 40 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 40 messages\n", - "\u001b[92m19:48:33 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/26dc2412-0699-4a4e-a272-dc576348a5c8/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 99%|███████████████████████████████████████-| 7270/7340 [262:15<2:31, 27.7 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/b4eee866-c191-4acf-b232-9b18a3c888ef/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:48:34,163 - agent.ComputerAgent - INFO - LLM processing started with 30 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 30 messages\n", - "\u001b[92m19:48:34 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 19:48:34,825 - agent.ComputerAgent - INFO - LLM processing started with 18 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 18 messages\n", - "\u001b[92m19:48:34 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 99%|███████████████████████████████████████-| 7270/7340 [262:16<2:31, 27.7 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/cb64a220-43d8-4373-bd2a-e73bacb4a122/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:48:36,005 - agent.ComputerAgent - INFO - LLM processing started with 26 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 26 messages\n", - "\u001b[92m19:48:36 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 99%|███████████████████████████████████████-| 7270/7340 [262:17<2:31, 27.7 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 99%|███████████████████████████████████████-| 7270/7340 [262:21<2:31, 27.7 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:48:40 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - " 99%|███████████████████████████████████████-| 7270/7340 [262:22<2:31, 27.7 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:48:41 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m19:48:41 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m19:48:42 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:48:43,277 - agent.ComputerAgent - INFO - Computer: click({'x': 66, 'y': 164})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 66, 'y': 164})\n", - " 99%|███████████████████████████████████████-| 7270/7340 [262:25<2:31, 27.7 steps/min]\u001b[92m19:48:43 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m19:48:43 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m19:48:44 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:48:44,471 - agent.ComputerAgent - INFO - Computer: click({'x': 248, 'y': 173})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 248, 'y': 173})\n", - " 99%|███████████████████████████████████████-| 7271/7340 [262:26<2:29, 27.7 steps/min]\u001b[92m19:48:44 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:48:45,143 - agent.ComputerAgent - INFO - Computer: drag({'path': [{'x': 288, 'y': 153}, {'x': 81, 'y': 155}]})\n", - "INFO:agent.ComputerAgent:Computer: drag({'path': [{'x': 288, 'y': 153}, {'x': 81, 'y': 155}]})\n", - " 99%|███████████████████████████████████████-| 7273/7340 [262:29<2:25, 27.7 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/d71be89e-00e2-40e7-8b8d-38e36bc6d26c/invoke \"HTTP/1.1 200 OK\"\n", - " 99%|███████████████████████████████████████-| 7273/7340 [262:30<2:25, 27.7 steps/min]2025-08-11 19:48:49,336 - agent.ComputerAgent - INFO - LLM processing started with 24 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 24 messages\n", - "\u001b[92m19:48:49 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 99%|███████████████████████████████████████-| 7273/7340 [262:31<2:25, 27.7 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/9053f5ae-149f-4a73-a89e-977f3e750435/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:48:50,547 - agent.ComputerAgent - INFO - LLM processing started with 18 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 18 messages\n", - "\u001b[92m19:48:50 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:48:51,906 - agent.ComputerAgent - INFO - Computer: type({'text': 'webui refresh 2023'})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': 'webui refresh 2023'})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/b14fe395-5fa2-43f0-9d0b-23c42f3e9093/invoke \"HTTP/1.1 200 OK\"\n", - " 99%|███████████████████████████████████████-| 7273/7340 [262:33<2:25, 27.7 steps/min]2025-08-11 19:48:52,576 - agent.ComputerAgent - INFO - LLM processing started with 22 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 22 messages\n", - "\u001b[92m19:48:52 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:48:53,923 - agent.ComputerAgent - INFO - Computer: type({'text': 'Thunderbird'})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': 'Thunderbird'})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - " 99%|███████████████████████████████████████-| 7274/7340 [262:36<2:22, 27.7 steps/min]\u001b[92m19:48:54 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m19:48:55 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:48:56,172 - agent.ComputerAgent - INFO - Computer: double_click({'x': 144, 'y': 167})\n", - "INFO:agent.ComputerAgent:Computer: double_click({'x': 144, 'y': 167})\n", - " 99%|███████████████████████████████████████-| 7276/7340 [262:39<2:18, 27.7 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:48:58 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - " 99%|███████████████████████████████████████-| 7276/7340 [262:40<2:18, 27.7 steps/min]\u001b[92m19:48:59 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:49:00,051 - agent.ComputerAgent - INFO - Computer: click({'x': 761, 'y': 229})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 761, 'y': 229})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/fa081188-4859-4858-9d33-0f9675111182/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/b4eee866-c191-4acf-b232-9b18a3c888ef/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:49:00,717 - agent.ComputerAgent - INFO - LLM processing started with 20 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 20 messages\n", - "\u001b[92m19:49:00 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 99%|███████████████████████████████████████-| 7276/7340 [262:42<2:18, 27.7 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:49:01 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 99%|███████████████████████████████████████-| 7277/7340 [262:43<2:16, 27.7 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m19:49:02 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m19:49:02 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/26dc2412-0699-4a4e-a272-dc576348a5c8/invoke \"HTTP/1.1 200 OK\"\n", - " 99%|███████████████████████████████████████-| 7277/7340 [262:44<2:16, 27.7 steps/min]2025-08-11 19:49:03,565 - agent.ComputerAgent - INFO - LLM processing started with 32 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 32 messages\n", - "\u001b[92m19:49:03 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m19:49:03 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:49:04,222 - agent.ComputerAgent - INFO - Computer: drag({'path': [{'x': 237, 'y': 178}, {'x': 258, 'y': 280}]})\n", - "INFO:agent.ComputerAgent:Computer: drag({'path': [{'x': 237, 'y': 178}, {'x': 258, 'y': 280}]})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/fa081188-4859-4858-9d33-0f9675111182/invoke \"HTTP/1.1 200 OK\"\n", - " 99%|███████████████████████████████████████-| 7277/7340 [262:45<2:16, 27.7 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/fa081188-4859-4858-9d33-0f9675111182/close \"HTTP/1.1 200 OK\"\n", - " 99%|███████████████████████████████████████-| 7278/7340 [262:46<2:14, 27.7 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/cb64a220-43d8-4373-bd2a-e73bacb4a122/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:49:06,401 - agent.ComputerAgent - INFO - LLM processing started with 28 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 28 messages\n", - "\u001b[92m19:49:06 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 99%|███████████████████████████████████████-| 7278/7340 [262:48<2:14, 27.7 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 99%|███████████████████████████████████████-| 7278/7340 [262:51<2:14, 27.7 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/9053f5ae-149f-4a73-a89e-977f3e750435/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:49:10,616 - agent.ComputerAgent - INFO - LLM processing started with 20 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 20 messages\n", - "\u001b[92m19:49:10 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 99%|███████████████████████████████████████-| 7278/7340 [262:52<2:14, 27.7 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 99%|███████████████████████████████████████-| 7278/7340 [262:53<2:14, 27.7 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:49:12 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m19:49:12 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - " 99%|███████████████████████████████████████-| 7278/7340 [262:55<2:14, 27.7 steps/min]\u001b[92m19:49:13 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m19:49:13 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:49:14,305 - agent.ComputerAgent - INFO - Computer: click({'x': 1008, 'y': 223})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 1008, 'y': 223})\n", - "\u001b[92m19:49:14 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:49:14,966 - agent.ComputerAgent - INFO - Computer: click({'x': 75, 'y': 135})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 75, 'y': 135})\n", - " 99%|███████████████████████████████████████-| 7278/7340 [262:56<2:14, 27.7 steps/min]\u001b[92m19:49:15 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:49:15,652 - agent.ComputerAgent - INFO - Computer: double_click({'x': 540, 'y': 128})\n", - "INFO:agent.ComputerAgent:Computer: double_click({'x': 540, 'y': 128})\n", - " 99%|███████████████████████████████████████-| 7281/7340 [262:58<2:07, 27.7 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:49:17 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - " 99%|███████████████████████████████████████-| 7281/7340 [262:59<2:07, 27.7 steps/min]\u001b[92m19:49:18 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:49:18,506 - agent.ComputerAgent - INFO - Computer: click({'x': 229, 'y': 157})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 229, 'y': 157})\n", - " 99%|███████████████████████████████████████-| 7281/7340 [263:00<2:07, 27.7 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/8e75deb1-3c97-408b-8c7d-f4681b322141/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:49:20,217 - agent.ComputerAgent - INFO - LLM processing started with 42 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 42 messages\n", - "\u001b[92m19:49:20 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 99%|███████████████████████████████████████-| 7282/7340 [263:02<2:05, 27.7 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:49:21 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/d71be89e-00e2-40e7-8b8d-38e36bc6d26c/invoke \"HTTP/1.1 200 OK\"\n", - " 99%|███████████████████████████████████████-| 7282/7340 [263:03<2:05, 27.7 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:49:23,091 - agent.ComputerAgent - INFO - LLM processing started with 26 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 26 messages\n", - "\u001b[92m19:49:23 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m19:49:23 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/b4eee866-c191-4acf-b232-9b18a3c888ef/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 99%|███████████████████████████████████████-| 7282/7340 [263:04<2:05, 27.7 steps/min]2025-08-11 19:49:23,787 - agent.ComputerAgent - INFO - Computer: click({'x': 237, 'y': 178})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 237, 'y': 178})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/b14fe395-5fa2-43f0-9d0b-23c42f3e9093/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:49:24,438 - agent.ComputerAgent - INFO - LLM processing started with 22 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 22 messages\n", - "\u001b[92m19:49:24 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 19:49:25,775 - agent.ComputerAgent - INFO - Computer: type({'text': \"=VLOOKUP(C2,'Retail Price'.$A$2:$B$200,2,0)*E2*(1-F2)\"})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': \"=VLOOKUP(C2,'Retail Price'.$A$2:$B$200,2,0)*E2*(1-F2)\"})\n", - " 99%|███████████████████████████████████████-| 7282/7340 [263:07<2:05, 27.7 steps/min]2025-08-11 19:49:26,452 - agent.ComputerAgent - INFO - LLM processing started with 24 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 24 messages\n", - "\u001b[92m19:49:26 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 99%|███████████████████████████████████████-| 7284/7340 [263:12<2:01, 27.7 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/9053f5ae-149f-4a73-a89e-977f3e750435/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/cb64a220-43d8-4373-bd2a-e73bacb4a122/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:49:31,690 - agent.ComputerAgent - INFO - LLM processing started with 22 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 22 messages\n", - "\u001b[92m19:49:31 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:49:32 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 99%|███████████████████████████████████████-| 7284/7340 [263:14<2:01, 27.7 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:49:33,032 - agent.ComputerAgent - INFO - LLM processing started with 30 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 30 messages\n", - "\u001b[92m19:49:33 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m19:49:33 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:49:33,738 - agent.ComputerAgent - INFO - Computer: click({'x': 1004, 'y': 60})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 1004, 'y': 60})\n", - " 99%|███████████████████████████████████████-| 7285/7340 [263:16<1:59, 27.7 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:49:35 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - " 99%|███████████████████████████████████████-| 7285/7340 [263:17<1:59, 27.7 steps/min]\u001b[92m19:49:36 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:49:36,569 - agent.ComputerAgent - INFO - Computer: click({'x': 980, 'y': 60})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 980, 'y': 60})\n", - " 99%|███████████████████████████████████████-| 7286/7340 [263:20<1:57, 27.7 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/26dc2412-0699-4a4e-a272-dc576348a5c8/invoke \"HTTP/1.1 200 OK\"\n", - " 99%|███████████████████████████████████████-| 7286/7340 [263:21<1:57, 27.7 steps/min]2025-08-11 19:49:40,273 - agent.ComputerAgent - INFO - LLM processing started with 34 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 34 messages\n", - "\u001b[92m19:49:40 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 99%|███████████████████████████████████████-| 7286/7340 [263:22<1:57, 27.7 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:49:42,639 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'CTRL+H'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'CTRL+H'})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/b4eee866-c191-4acf-b232-9b18a3c888ef/invoke \"HTTP/1.1 200 OK\"\n", - " 99%|███████████████████████████████████████-| 7286/7340 [263:24<1:57, 27.7 steps/min]2025-08-11 19:49:43,312 - agent.ComputerAgent - INFO - LLM processing started with 24 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 24 messages\n", - "\u001b[92m19:49:43 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 99%|███████████████████████████████████████-| 7286/7340 [263:25<1:57, 27.7 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:49:44 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m19:49:45 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 99%|███████████████████████████████████████-| 7286/7340 [263:26<1:57, 27.7 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m19:49:45 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m19:49:45 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - " 99%|███████████████████████████████████████-| 7286/7340 [263:27<1:57, 27.7 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:49:47,535 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'enter'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'enter'})\n", - "\u001b[92m19:49:47 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/8e75deb1-3c97-408b-8c7d-f4681b322141/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m19:49:47 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:49:48,871 - agent.ComputerAgent - INFO - Computer: screenshot({})\n", - "INFO:agent.ComputerAgent:Computer: screenshot({})\n", - " 99%|███████████████████████████████████████-| 7286/7340 [263:30<1:57, 27.6 steps/min]2025-08-11 19:49:49,547 - agent.ComputerAgent - INFO - Computer: click({'x': 259, 'y': 180})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 259, 'y': 180})\n", - "2025-08-11 19:49:50,246 - agent.ComputerAgent - INFO - Computer: drag({'path': [{'x': 55, 'y': 164}, {'x': 55, 'y': 600}]})\n", - "INFO:agent.ComputerAgent:Computer: drag({'path': [{'x': 55, 'y': 164}, {'x': 55, 'y': 600}]})\n", - " 99%|███████████████████████████████████████-| 7291/7340 [263:33<1:46, 27.7 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/8e75deb1-3c97-408b-8c7d-f4681b322141/close \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - " 99%|███████████████████████████████████████-| 7291/7340 [263:34<1:46, 27.7 steps/min]\u001b[92m19:49:52 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/b14fe395-5fa2-43f0-9d0b-23c42f3e9093/invoke \"HTTP/1.1 200 OK\"\n" - ] - }, - { - "name": "stdout", - "output_type": "stream", - "text": [ - "No screenshot found, taking screenshot\n" - ] - }, - { - "name": "stderr", - "output_type": "stream", - "text": [ - "2025-08-11 19:49:53,552 - agent.ComputerAgent - INFO - LLM processing started with 26 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 26 messages\n", - "\u001b[92m19:49:53 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m19:49:53 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:49:54,210 - agent.ComputerAgent - INFO - Computer: click({'x': 867, 'y': 296})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 867, 'y': 296})\n", - " 99%|███████████████████████████████████████-| 7291/7340 [263:35<1:46, 27.7 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/9053f5ae-149f-4a73-a89e-977f3e750435/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/cb64a220-43d8-4373-bd2a-e73bacb4a122/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:49:55,903 - agent.ComputerAgent - INFO - LLM processing started with 24 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 24 messages\n", - "\u001b[92m19:49:55 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 99%|███████████████████████████████████████-| 7292/7340 [263:37<1:44, 27.7 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/d71be89e-00e2-40e7-8b8d-38e36bc6d26c/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:49:56,562 - agent.ComputerAgent - INFO - LLM processing started with 32 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 32 messages\n", - "\u001b[92m19:49:56 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 19:49:57,261 - agent.ComputerAgent - INFO - LLM processing started with 28 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 28 messages\n", - "\u001b[92m19:49:57 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 99%|███████████████████████████████████████-| 7292/7340 [263:39<1:44, 27.7 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 99%|███████████████████████████████████████-| 7292/7340 [263:41<1:44, 27.7 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/b4eee866-c191-4acf-b232-9b18a3c888ef/invoke \"HTTP/1.1 200 OK\"\n", - " 99%|███████████████████████████████████████-| 7292/7340 [263:42<1:44, 27.7 steps/min]2025-08-11 19:50:00,993 - agent.ComputerAgent - INFO - LLM processing started with 26 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 26 messages\n", - "\u001b[92m19:50:01 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 99%|███████████████████████████████████████-| 7292/7340 [263:49<1:44, 27.6 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:50:08,913 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'ctrl+c'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'ctrl+c'})\n", - " 99%|███████████████████████████████████████-| 7292/7340 [263:50<1:44, 27.6 steps/min]2025-08-11 19:50:10,074 - agent.ComputerAgent - INFO - LLM processing started with 26 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 26 messages\n", - "\u001b[92m19:50:10 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 99%|███████████████████████████████████████-| 7292/7340 [263:51<1:44, 27.6 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 99%|███████████████████████████████████████-| 7292/7340 [263:52<1:44, 27.6 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:50:12 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 99%|███████████████████████████████████████-| 7292/7340 [263:54<1:44, 27.6 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:50:13 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 99%|███████████████████████████████████████-| 7292/7340 [263:55<1:44, 27.6 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m19:50:13 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:50:14,024 - agent.ComputerAgent - INFO - Computer: click({'x': 871, 'y': 135})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 871, 'y': 135})\n", - "\u001b[92m19:50:14 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m19:50:14 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - " 99%|███████████████████████████████████████-| 7292/7340 [263:56<1:44, 27.6 steps/min]\u001b[92m19:50:14 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:50:15,204 - agent.ComputerAgent - INFO - Computer: drag({'path': [{'x': 275, 'y': 152}, {'x': 79, 'y': 154}]})\n", - "INFO:agent.ComputerAgent:Computer: drag({'path': [{'x': 275, 'y': 152}, {'x': 79, 'y': 154}]})\n", - " 99%|███████████████████████████████████████-| 7293/7340 [263:57<1:42, 27.6 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:50:17,610 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'ALT+TAB'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'ALT+TAB'})\n", - " 99%|███████████████████████████████████████-| 7294/7340 [263:59<1:39, 27.6 steps/min]2025-08-11 19:50:18,781 - agent.ComputerAgent - INFO - LLM processing started with 36 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 36 messages\n", - "\u001b[92m19:50:18 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 99%|███████████████████████████████████████-| 7294/7340 [264:00<1:39, 27.6 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 99%|███████████████████████████████████████-| 7294/7340 [264:01<1:39, 27.6 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/b4eee866-c191-4acf-b232-9b18a3c888ef/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:50:20 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:50:21,132 - agent.ComputerAgent - INFO - LLM processing started with 28 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 28 messages\n", - "\u001b[92m19:50:21 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/b14fe395-5fa2-43f0-9d0b-23c42f3e9093/invoke \"HTTP/1.1 200 OK\"\n", - " 99%|███████████████████████████████████████-| 7294/7340 [264:02<1:39, 27.6 steps/min]2025-08-11 19:50:22,169 - agent.ComputerAgent - INFO - LLM processing started with 28 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 28 messages\n", - "\u001b[92m19:50:22 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 99%|███████████████████████████████████████-| 7294/7340 [264:03<1:39, 27.6 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m19:50:22 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:50:22,830 - agent.ComputerAgent - INFO - Computer: click({'x': 120, 'y': 53})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 120, 'y': 53})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:50:24,191 - agent.ComputerAgent - INFO - Computer: drag({'path': [{'x': 257, 'y': 153}, {'x': 259, 'y': 281}]})\n", - "INFO:agent.ComputerAgent:Computer: drag({'path': [{'x': 257, 'y': 153}, {'x': 259, 'y': 281}]})\n", - " 99%|███████████████████████████████████████-| 7296/7340 [264:08<1:35, 27.6 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:50:27 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - " 99%|███████████████████████████████████████-| 7296/7340 [264:09<1:35, 27.6 steps/min]\u001b[92m19:50:28 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:50:29,058 - agent.ComputerAgent - INFO - Computer: click({'x': 749, 'y': 229})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 749, 'y': 229})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/d71be89e-00e2-40e7-8b8d-38e36bc6d26c/invoke \"HTTP/1.1 200 OK\"\n", - " 99%|███████████████████████████████████████-| 7296/7340 [264:10<1:35, 27.6 steps/min]2025-08-11 19:50:29,733 - agent.ComputerAgent - INFO - LLM processing started with 30 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 30 messages\n", - "\u001b[92m19:50:29 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/9053f5ae-149f-4a73-a89e-977f3e750435/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 19:50:30,431 - agent.ComputerAgent - INFO - LLM processing started with 28 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 28 messages\n", - "\u001b[92m19:50:30 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 99%|███████████████████████████████████████-| 7297/7340 [264:12<1:33, 27.6 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 99%|███████████████████████████████████████-| 7297/7340 [264:16<1:33, 27.6 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:50:35 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/cb64a220-43d8-4373-bd2a-e73bacb4a122/invoke \"HTTP/1.1 200 OK\"\n", - " 99%|███████████████████████████████████████-| 7297/7340 [264:17<1:33, 27.6 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:50:36,275 - agent.ComputerAgent - INFO - LLM processing started with 34 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 34 messages\n", - "\u001b[92m19:50:36 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m19:50:36 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:50:37,340 - agent.ComputerAgent - INFO - Computer: click({'x': 229, 'y': 91})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 229, 'y': 91})\n", - " 99%|███████████████████████████████████████-| 7298/7340 [264:21<1:31, 27.6 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:50:41,280 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'ALT+TAB'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'ALT+TAB'})\n", - " 99%|███████████████████████████████████████-| 7298/7340 [264:23<1:31, 27.6 steps/min]2025-08-11 19:50:42,453 - agent.ComputerAgent - INFO - LLM processing started with 38 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 38 messages\n", - "\u001b[92m19:50:42 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 99%|███████████████████████████████████████-| 7298/7340 [264:24<1:31, 27.6 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 99%|███████████████████████████████████████-| 7298/7340 [264:25<1:31, 27.6 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/b4eee866-c191-4acf-b232-9b18a3c888ef/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:50:44,654 - agent.ComputerAgent - INFO - LLM processing started with 30 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 30 messages\n", - "\u001b[92m19:50:44 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 99%|███████████████████████████████████████-| 7298/7340 [264:26<1:31, 27.6 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 99%|███████████████████████████████████████-| 7298/7340 [264:28<1:31, 27.6 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:50:47 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - " 99%|███████████████████████████████████████-| 7298/7340 [264:29<1:31, 27.6 steps/min]\u001b[92m19:50:48 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:50:48,511 - agent.ComputerAgent - INFO - Computer: click({'x': 243, 'y': 178})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 243, 'y': 178})\n", - " 99%|███████████████████████████████████████-| 7298/7340 [264:30<1:31, 27.6 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:50:49 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 99%|███████████████████████████████████████-| 7299/7340 [264:31<1:29, 27.6 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m19:50:50 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:50:50,849 - agent.ComputerAgent - INFO - Computer: click({'x': 138, 'y': 90})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 138, 'y': 90})\n", - " 99%|███████████████████████████████████████-| 7299/7340 [264:32<1:29, 27.6 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:50:52 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - " 99%|███████████████████████████████████████-| 7300/7340 [264:33<1:26, 27.6 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m19:50:52 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m19:50:52 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - " 99%|███████████████████████████████████████-| 7300/7340 [264:34<1:26, 27.6 steps/min]\u001b[92m19:50:53 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:50:53,680 - agent.ComputerAgent - INFO - Computer: drag({'path': [{'x': 749, 'y': 183}, {'x': 837, 'y': 244}]})\n", - "INFO:agent.ComputerAgent:Computer: drag({'path': [{'x': 749, 'y': 183}, {'x': 837, 'y': 244}]})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/9053f5ae-149f-4a73-a89e-977f3e750435/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:50:54 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - " 99%|███████████████████████████████████████-| 7300/7340 [264:36<1:26, 27.6 steps/min]\u001b[92m19:50:55 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:50:55,684 - agent.ComputerAgent - INFO - LLM processing started with 30 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 30 messages\n", - "\u001b[92m19:50:55 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m19:50:55 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:50:56,394 - agent.ComputerAgent - INFO - Computer: click({'x': 258, 'y': 155})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 258, 'y': 155})\n", - " 99%|███████████████████████████████████████-| 7301/7340 [264:38<1:24, 27.6 steps/min]\u001b[92m19:50:56 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:50:57,053 - agent.ComputerAgent - INFO - Computer: click({'button': 'right', 'x': 118, 'y': 182})\n", - "INFO:agent.ComputerAgent:Computer: click({'button': 'right', 'x': 118, 'y': 182})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/d71be89e-00e2-40e7-8b8d-38e36bc6d26c/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:50:57,705 - agent.ComputerAgent - INFO - LLM processing started with 32 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 32 messages\n", - "\u001b[92m19:50:57 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 99%|███████████████████████████████████████-| 7302/7340 [264:39<1:22, 27.6 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 99%|███████████████████████████████████████-| 7303/7340 [264:42<1:20, 27.6 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/cb64a220-43d8-4373-bd2a-e73bacb4a122/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:51:01,944 - agent.ComputerAgent - INFO - LLM processing started with 36 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 36 messages\n", - "\u001b[92m19:51:01 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/b14fe395-5fa2-43f0-9d0b-23c42f3e9093/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 99%|███████████████████████████████████████-| 7303/7340 [264:43<1:20, 27.6 steps/min]2025-08-11 19:51:02,644 - agent.ComputerAgent - INFO - LLM processing started with 30 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 30 messages\n", - "\u001b[92m19:51:02 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 99%|███████████████████████████████████████-| 7303/7340 [264:44<1:20, 27.6 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/b4eee866-c191-4acf-b232-9b18a3c888ef/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:51:04,313 - agent.ComputerAgent - INFO - LLM processing started with 32 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 32 messages\n", - "\u001b[92m19:51:04 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - " 99%|███████████████████████████████████████-| 7303/7340 [264:46<1:20, 27.6 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:51:05 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - " 99%|███████████████████████████████████████-| 7303/7340 [264:47<1:20, 27.6 steps/min]\u001b[92m19:51:05 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:51:06,170 - agent.ComputerAgent - INFO - Computer: click({'x': 285, 'y': 98})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 285, 'y': 98})\n", - "100%|███████████████████████████████████████-| 7304/7340 [264:53<1:18, 27.6 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/26dc2412-0699-4a4e-a272-dc576348a5c8/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:51:12,936 - agent.ComputerAgent - INFO - LLM processing started with 40 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 40 messages\n", - "\u001b[92m19:51:13 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "100%|███████████████████████████████████████-| 7304/7340 [264:54<1:18, 27.6 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "100%|███████████████████████████████████████-| 7304/7340 [264:57<1:18, 27.6 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:51:16 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m19:51:17 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "100%|███████████████████████████████████████-| 7304/7340 [264:59<1:18, 27.6 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m19:51:17 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:51:18,032 - agent.ComputerAgent - INFO - Computer: click({'x': 259, 'y': 180})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 259, 'y': 180})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:51:18 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "100%|███████████████████████████████████████-| 7304/7340 [265:00<1:18, 27.6 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m19:51:18 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m19:51:18 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m19:51:19 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:51:19,864 - agent.ComputerAgent - INFO - Computer: click({'x': 151, 'y': 232})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 151, 'y': 232})\n", - "100%|███████████████████████████████████████-| 7305/7340 [265:01<1:16, 27.6 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:51:20 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m19:51:20 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:51:21,195 - agent.ComputerAgent - INFO - Computer: drag({'path': [{'x': 298, 'y': 152}, {'x': 81, 'y': 155}]})\n", - "INFO:agent.ComputerAgent:Computer: drag({'path': [{'x': 298, 'y': 152}, {'x': 81, 'y': 155}]})\n", - "100%|███████████████████████████████████████-| 7306/7340 [265:02<1:14, 27.6 steps/min]\u001b[92m19:51:21 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:51:21,882 - agent.ComputerAgent - INFO - Computer: click({'x': 59, 'y': 157})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 59, 'y': 157})\n", - "100%|███████████████████████████████████████-| 7307/7340 [265:03<1:11, 27.6 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/9053f5ae-149f-4a73-a89e-977f3e750435/invoke \"HTTP/1.1 200 OK\"\n", - "100%|███████████████████████████████████████-| 7308/7340 [265:04<1:09, 27.6 steps/min]2025-08-11 19:51:23,543 - agent.ComputerAgent - INFO - LLM processing started with 32 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 32 messages\n", - "\u001b[92m19:51:23 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "100%|███████████████████████████████████████-| 7308/7340 [265:06<1:09, 27.6 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:51:25 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "100%|███████████████████████████████████████-| 7308/7340 [265:07<1:09, 27.6 steps/min]\u001b[92m19:51:26 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:51:26,961 - agent.ComputerAgent - INFO - Computer: click({'x': 762, 'y': 230})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 762, 'y': 230})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/b14fe395-5fa2-43f0-9d0b-23c42f3e9093/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:51:27,615 - agent.ComputerAgent - INFO - LLM processing started with 32 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 32 messages\n", - "\u001b[92m19:51:27 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/b4eee866-c191-4acf-b232-9b18a3c888ef/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "100%|███████████████████████████████████████-| 7308/7340 [265:09<1:09, 27.6 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/d71be89e-00e2-40e7-8b8d-38e36bc6d26c/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:51:28,303 - agent.ComputerAgent - INFO - LLM processing started with 34 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 34 messages\n", - "\u001b[92m19:51:28 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 19:51:28,994 - agent.ComputerAgent - INFO - LLM processing started with 34 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 34 messages\n", - "\u001b[92m19:51:29 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "100%|███████████████████████████████████████-| 7309/7340 [265:15<1:07, 27.6 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/cb64a220-43d8-4373-bd2a-e73bacb4a122/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:51:35,225 - agent.ComputerAgent - INFO - LLM processing started with 38 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 38 messages\n", - "\u001b[92m19:51:35 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "100%|███████████████████████████████████████-| 7309/7340 [265:17<1:07, 27.6 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "100%|███████████████████████████████████████-| 7309/7340 [265:18<1:07, 27.5 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:51:38,065 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'ALT+TAB'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'ALT+TAB'})\n", - "100%|███████████████████████████████████████-| 7309/7340 [265:19<1:07, 27.5 steps/min]2025-08-11 19:51:39,205 - agent.ComputerAgent - INFO - LLM processing started with 42 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 42 messages\n", - "\u001b[92m19:51:39 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "100%|███████████████████████████████████████-| 7309/7340 [265:21<1:07, 27.5 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "100%|███████████████████████████████████████-| 7309/7340 [265:22<1:07, 27.5 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:51:40 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "100%|███████████████████████████████████████-| 7309/7340 [265:23<1:07, 27.5 steps/min]\u001b[92m19:51:41 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:51:42,081 - agent.ComputerAgent - INFO - Computer: click({'x': 237, 'y': 95})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 237, 'y': 95})\n", - "100%|███████████████████████████████████████-| 7310/7340 [265:25<1:05, 27.5 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:51:44 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "100%|███████████████████████████████████████-| 7310/7340 [265:26<1:05, 27.5 steps/min]\u001b[92m19:51:44 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m19:51:45 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "100%|███████████████████████████████████████-| 7310/7340 [265:27<1:05, 27.5 steps/min]\u001b[92m19:51:45 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m19:51:45 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m19:51:46 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:51:46,954 - agent.ComputerAgent - INFO - Computer: drag({'start_element_description': 'Cell C2', 'end_element_description': 'Cell C10', 'x': 268, 'y': 188})\n", - "INFO:agent.ComputerAgent:Computer: drag({'start_element_description': 'Cell C2', 'end_element_description': 'Cell C10', 'x': 268, 'y': 188})\n", - "100%|███████████████████████████████████████-| 7311/7340 [265:29<1:03, 27.5 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/9053f5ae-149f-4a73-a89e-977f3e750435/invoke \"HTTP/1.1 200 OK\"\n" - ] - }, - { - "name": "stdout", - "output_type": "stream", - "text": [ - "No screenshot found, taking screenshot\n" - ] - }, - { - "name": "stderr", - "output_type": "stream", - "text": [ - "2025-08-11 19:51:48,615 - agent.ComputerAgent - INFO - LLM processing started with 35 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 35 messages\n", - "\u001b[92m19:51:48 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/b4eee866-c191-4acf-b232-9b18a3c888ef/invoke \"HTTP/1.1 200 OK\"\n", - "100%|███████████████████████████████████████-| 7311/7340 [265:30<1:03, 27.5 steps/min]2025-08-11 19:51:49,299 - agent.ComputerAgent - INFO - LLM processing started with 36 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 36 messages\n", - "\u001b[92m19:51:49 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "100%|███████████████████████████████████████-| 7311/7340 [265:31<1:03, 27.5 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:51:51 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "100%|███████████████████████████████████████-| 7311/7340 [265:32<1:03, 27.5 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 400 Bad Request\"\n", - "\u001b[92m19:51:51 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:51:52,164 - agent.ComputerAgent - INFO - Computer: click({'x': 87, 'y': 166})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 87, 'y': 166})\n", - "100%|███████████████████████████████████████-| 7312/7340 [265:33<1:01, 27.5 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/9053f5ae-149f-4a73-a89e-977f3e750435/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:51:52,805 - agent.ComputerAgent - INFO - LLM processing started with 37 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 37 messages\n", - "\u001b[92m19:51:52 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "100%|███████████████████████████████████████-| 7313/7340 [265:35<0:58, 27.5 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:51:54 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "100%|███████████████████████████████████████-| 7313/7340 [265:36<0:58, 27.5 steps/min]\u001b[92m19:51:55 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:51:55,701 - agent.ComputerAgent - INFO - Computer: double_click({'x': 841, 'y': 244})\n", - "INFO:agent.ComputerAgent:Computer: double_click({'x': 841, 'y': 244})\n", - "100%|███████████████████████████████████████-| 7313/7340 [265:37<0:58, 27.5 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 400 Bad Request\"\n", - "100%|███████████████████████████████████████-| 7315/7340 [265:38<0:54, 27.5 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/b14fe395-5fa2-43f0-9d0b-23c42f3e9093/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/9053f5ae-149f-4a73-a89e-977f3e750435/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:51:57,894 - agent.ComputerAgent - INFO - LLM processing started with 39 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 39 messages\n", - "\u001b[92m19:51:57 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "2025-08-11 19:51:58,566 - agent.ComputerAgent - INFO - LLM processing started with 34 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 34 messages\n", - "\u001b[92m19:51:58 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "100%|███████████████████████████████████████-| 7315/7340 [265:40<0:54, 27.5 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:51:59 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "100%|███████████████████████████████████████-| 7315/7340 [265:41<0:54, 27.5 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m19:52:00 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:52:00,770 - agent.ComputerAgent - INFO - Computer: click({'x': 422, 'y': 360})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 422, 'y': 360})\n", - "100%|███████████████████████████████████████-| 7315/7340 [265:42<0:54, 27.5 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/cb64a220-43d8-4373-bd2a-e73bacb4a122/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:52:01,926 - agent.ComputerAgent - INFO - LLM processing started with 40 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 40 messages\n", - "\u001b[92m19:52:01 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "100%|███████████████████████████████████████-| 7316/7340 [265:43<0:52, 27.5 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 400 Bad Request\"\n", - "100%|███████████████████████████████████████-| 7317/7340 [265:44<0:50, 27.5 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:52:03 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "100%|███████████████████████████████████████-| 7317/7340 [265:45<0:50, 27.5 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/9053f5ae-149f-4a73-a89e-977f3e750435/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:52:04,815 - agent.ComputerAgent - INFO - LLM processing started with 41 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 41 messages\n", - "\u001b[92m19:52:04 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m19:52:04 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:52:05,539 - agent.ComputerAgent - INFO - Computer: click({'x': 15, 'y': 526})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 15, 'y': 526})\n", - "100%|███████████████████████████████████████-| 7317/7340 [265:47<0:50, 27.5 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/d71be89e-00e2-40e7-8b8d-38e36bc6d26c/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:52:07,328 - agent.ComputerAgent - INFO - Computer: type({'text': 'contacts.csv'})\n", - "INFO:agent.ComputerAgent:Computer: type({'text': 'contacts.csv'})\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 400 Bad Request\"\n", - "100%|███████████████████████████████████████-| 7319/7340 [265:49<0:45, 27.5 steps/min]2025-08-11 19:52:07,988 - agent.ComputerAgent - INFO - LLM processing started with 36 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 36 messages\n", - "\u001b[92m19:52:08 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "100%|███████████████████████████████████████-| 7320/7340 [265:50<0:43, 27.5 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/9053f5ae-149f-4a73-a89e-977f3e750435/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:52:09,185 - agent.ComputerAgent - INFO - LLM processing started with 43 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 43 messages\n", - "\u001b[92m19:52:09 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "100%|███████████████████████████████████████-| 7320/7340 [265:53<0:43, 27.5 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/26dc2412-0699-4a4e-a272-dc576348a5c8/invoke \"HTTP/1.1 200 OK\"\n", - "100%|███████████████████████████████████████-| 7320/7340 [265:54<0:43, 27.5 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 400 Bad Request\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/b4eee866-c191-4acf-b232-9b18a3c888ef/invoke \"HTTP/1.1 200 OK\"\n", - "100%|███████████████████████████████████████-| 7321/7340 [265:55<0:41, 27.5 steps/min]2025-08-11 19:52:13,898 - agent.ComputerAgent - INFO - LLM processing started with 38 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 38 messages\n", - "\u001b[92m19:52:13 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/9053f5ae-149f-4a73-a89e-977f3e750435/invoke \"HTTP/1.1 200 OK\"\n", - "100%|███████████████████████████████████████-| 7321/7340 [265:58<0:41, 27.5 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:52:18 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/9053f5ae-149f-4a73-a89e-977f3e750435/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "100%|███████████████████████████████████████-| 7321/7340 [265:59<0:41, 27.5 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/26dc2412-0699-4a4e-a272-dc576348a5c8/invoke \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:52:18 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m19:52:18 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/9053f5ae-149f-4a73-a89e-977f3e750435/close \"HTTP/1.1 200 OK\"\n", - "100%|███████████████████████████████████████-| 7325/7340 [266:00<0:32, 27.5 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/26dc2412-0699-4a4e-a272-dc576348a5c8/close \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:52:19 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m19:52:19 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "100%|███████████████████████████████████████-| 7325/7340 [266:01<0:32, 27.5 steps/min]\u001b[92m19:52:20 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:52:20,887 - agent.ComputerAgent - INFO - Computer: drag({'start_element_description': 'column header A', 'end_element_description': 'column header H', 'x': 90, 'y': 155})\n", - "INFO:agent.ComputerAgent:Computer: drag({'start_element_description': 'column header A', 'end_element_description': 'column header H', 'x': 90, 'y': 155})\n", - "100%|███████████████████████████████████████-| 7326/7340 [266:02<0:30, 27.5 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:52:23,480 - agent.ComputerAgent - INFO - Computer: keypress({'keys': 'enter'})\n", - "INFO:agent.ComputerAgent:Computer: keypress({'keys': 'enter'})\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/b14fe395-5fa2-43f0-9d0b-23c42f3e9093/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "100%|███████████████████████████████████████-| 7326/7340 [266:05<0:30, 27.5 steps/min]\u001b[92m19:52:24 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n" - ] - }, - { - "name": "stdout", - "output_type": "stream", - "text": [ - "No screenshot found, taking screenshot\n" - ] - }, - { - "name": "stderr", - "output_type": "stream", - "text": [ - "2025-08-11 19:52:24,777 - agent.ComputerAgent - INFO - LLM processing started with 37 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 37 messages\n", - "\u001b[92m19:52:24 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "100%|███████████████████████████████████████-| 7327/7340 [266:06<0:28, 27.5 steps/min]\u001b[92m19:52:25 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:52:26,307 - agent.ComputerAgent - INFO - Computer: click({'x': 828, 'y': 35})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 828, 'y': 35})\n", - "100%|███████████████████████████████████████-| 7328/7340 [266:10<0:26, 27.5 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 400 Bad Request\"\n", - "100%|███████████████████████████████████████-| 7329/7340 [266:11<0:23, 27.5 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/cb64a220-43d8-4373-bd2a-e73bacb4a122/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:52:30,881 - agent.ComputerAgent - INFO - LLM processing started with 42 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 42 messages\n", - "\u001b[92m19:52:30 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/b14fe395-5fa2-43f0-9d0b-23c42f3e9093/invoke \"HTTP/1.1 200 OK\"\n", - "100%|███████████████████████████████████████-| 7329/7340 [266:12<0:23, 27.5 steps/min]2025-08-11 19:52:31,557 - agent.ComputerAgent - INFO - LLM processing started with 39 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 39 messages\n", - "\u001b[92m19:52:31 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "100%|███████████████████████████████████████-| 7329/7340 [266:13<0:23, 27.5 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/b4eee866-c191-4acf-b232-9b18a3c888ef/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:52:32,766 - agent.ComputerAgent - INFO - LLM processing started with 40 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 40 messages\n", - "\u001b[92m19:52:32 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "100%|███████████████████████████████████████-| 7329/7340 [266:19<0:23, 27.5 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 400 Bad Request\"\n", - "100%|███████████████████████████████████████-| 7330/7340 [266:20<0:21, 27.5 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:52:40 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/b14fe395-5fa2-43f0-9d0b-23c42f3e9093/invoke \"HTTP/1.1 200 OK\"\n", - "100%|███████████████████████████████████████-| 7330/7340 [266:21<0:21, 27.5 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:52:40,663 - agent.ComputerAgent - INFO - LLM processing started with 41 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 41 messages\n", - "\u001b[92m19:52:40 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m19:52:40 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:52:41,739 - agent.ComputerAgent - INFO - Computer: click({'x': 328, 'y': 286})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 328, 'y': 286})\n", - "100%|███████████████████████████████████████-| 7331/7340 [266:26<0:19, 27.5 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 400 Bad Request\"\n", - "100%|███████████████████████████████████████-| 7332/7340 [266:27<0:17, 27.5 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/b14fe395-5fa2-43f0-9d0b-23c42f3e9093/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:52:46,968 - agent.ComputerAgent - INFO - LLM processing started with 43 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 43 messages\n", - "\u001b[92m19:52:47 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "100%|███████████████████████████████████████-| 7332/7340 [266:29<0:17, 27.5 steps/min]\u001b[92m19:52:47 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/d71be89e-00e2-40e7-8b8d-38e36bc6d26c/invoke \"HTTP/1.1 200 OK\"\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "2025-08-11 19:52:48,348 - agent.ComputerAgent - INFO - LLM processing started with 38 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 38 messages\n", - "\u001b[92m19:52:48 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "\u001b[92m19:52:48 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "\u001b[92m19:52:48 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "100%|███████████████████████████████████████-| 7332/7340 [266:30<0:17, 27.5 steps/min]\u001b[92m19:52:49 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:52:49,902 - agent.ComputerAgent - INFO - Computer: drag({'path': [{'x': 749, 'y': 229}, {'x': 749, 'y': 732}]})\n", - "INFO:agent.ComputerAgent:Computer: drag({'path': [{'x': 749, 'y': 229}, {'x': 749, 'y': 732}]})\n", - "100%|███████████████████████████████████████-| 7332/7340 [266:31<0:17, 27.5 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:52:51 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 400 Bad Request\"\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "100%|███████████████████████████████████████-| 7334/7340 [266:32<0:13, 27.5 steps/min]Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m19:52:51 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:52:52,289 - agent.ComputerAgent - INFO - Computer: click({'x': 17, 'y': 386})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 17, 'y': 386})\n", - "100%|███████████████████████████████████████-| 7334/7340 [266:34<0:13, 27.5 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/b14fe395-5fa2-43f0-9d0b-23c42f3e9093/invoke \"HTTP/1.1 200 OK\"\n", - "100%|███████████████████████████████████████-| 7335/7340 [266:37<0:10, 27.5 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/cb64a220-43d8-4373-bd2a-e73bacb4a122/invoke \"HTTP/1.1 200 OK\"\n", - "100%|███████████████████████████████████████-| 7335/7340 [266:38<0:10, 27.5 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/b14fe395-5fa2-43f0-9d0b-23c42f3e9093/invoke \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/b14fe395-5fa2-43f0-9d0b-23c42f3e9093/close \"HTTP/1.1 200 OK\"\n", - "100%|███████████████████████████████████████-| 7335/7340 [266:39<0:10, 27.5 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/b4eee866-c191-4acf-b232-9b18a3c888ef/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:52:58,539 - agent.ComputerAgent - INFO - LLM processing started with 42 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 42 messages\n", - "\u001b[92m19:52:58 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "100%|███████████████████████████████████████-| 7335/7340 [266:40<0:10, 27.5 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/cb64a220-43d8-4373-bd2a-e73bacb4a122/invoke \"HTTP/1.1 200 OK\"\n", - "100%|███████████████████████████████████████-| 7335/7340 [266:41<0:10, 27.5 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/cb64a220-43d8-4373-bd2a-e73bacb4a122/close \"HTTP/1.1 200 OK\"\n", - "100%|███████████████████████████████████████-| 7335/7340 [266:54<0:10, 27.5 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:53:13 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "100%|███████████████████████████████████████-| 7335/7340 [266:55<0:10, 27.5 steps/min]\u001b[92m19:53:14 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:53:14,687 - agent.ComputerAgent - INFO - Computer: click({'x': 318, 'y': 306})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 318, 'y': 306})\n", - "100%|███████████████████████████████████████-| 7336/7340 [266:57<0:08, 27.5 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:53:16 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "100%|███████████████████████████████████████-| 7336/7340 [266:58<0:08, 27.5 steps/min]\u001b[92m19:53:16 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:53:17,018 - agent.ComputerAgent - INFO - Computer: click({'x': 49, 'y': 53})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 49, 'y': 53})\n", - "100%|███████████████████████████████████████-| 7337/7340 [267:01<0:06, 27.5 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/d71be89e-00e2-40e7-8b8d-38e36bc6d26c/invoke \"HTTP/1.1 200 OK\"\n", - "2025-08-11 19:53:20,724 - agent.ComputerAgent - INFO - LLM processing started with 40 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 40 messages\n", - "\u001b[92m19:53:20 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "100%|███████████████████████████████████████-| 7337/7340 [267:02<0:06, 27.5 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "100%|███████████████████████████████████████-| 7337/7340 [267:03<0:06, 27.5 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/b4eee866-c191-4acf-b232-9b18a3c888ef/invoke \"HTTP/1.1 200 OK\"\n", - "100%|███████████████████████████████████████-| 7337/7340 [267:23<0:06, 27.4 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:53:42 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "100%|███████████████████████████████████████-| 7337/7340 [267:24<0:06, 27.4 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/b4eee866-c191-4acf-b232-9b18a3c888ef/invoke \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:53:43 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:53:43,990 - agent.ComputerAgent - INFO - Computer: click({'x': 432, 'y': 314})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 432, 'y': 314})\n", - "100%|███████████████████████████████████████-| 7337/7340 [267:25<0:06, 27.4 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/b4eee866-c191-4acf-b232-9b18a3c888ef/close \"HTTP/1.1 200 OK\"\n", - "100%|███████████████████████████████████████-| 7338/7340 [267:29<0:04, 27.4 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/d71be89e-00e2-40e7-8b8d-38e36bc6d26c/invoke \"HTTP/1.1 200 OK\"\n", - "100%|███████████████████████████████████████-| 7338/7340 [267:30<0:04, 27.4 steps/min]2025-08-11 19:53:49,710 - agent.ComputerAgent - INFO - LLM processing started with 42 messages\n", - "INFO:agent.ComputerAgent:LLM processing started with 42 messages\n", - "\u001b[92m19:53:49 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= gpt-5; provider = openai\n", - "100%|███████████████████████████████████████-| 7338/7340 [268:00<0:04, 27.4 steps/min]INFO:httpx:HTTP Request: POST https://api.openai.com/v1/chat/completions \"HTTP/1.1 200 OK\"\n", - "\u001b[92m19:54:20 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "100%|███████████████████████████████████████-| 7338/7340 [268:02<0:04, 27.4 steps/min]INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "Setting `pad_token_id` to `eos_token_id`:151645 for open-end generation.\n", - "\u001b[92m19:54:21 - LiteLLM:INFO\u001b[0m: utils.py:3258 - \n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "INFO:LiteLLM:\n", - "LiteLLM completion() model= HelloKKMe/GTA1-7B; provider = huggingface-local\n", - "2025-08-11 19:54:21,613 - agent.ComputerAgent - INFO - Computer: click({'x': 469, 'y': 487})\n", - "INFO:agent.ComputerAgent:Computer: click({'x': 469, 'y': 487})\n", - "100%|███████████████████████████████████████-| 7339/7340 [268:08<0:02, 27.4 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/d71be89e-00e2-40e7-8b8d-38e36bc6d26c/invoke \"HTTP/1.1 200 OK\"\n", - "100%|███████████████████████████████████████-| 7339/7340 [268:12<0:02, 27.4 steps/min]INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/d71be89e-00e2-40e7-8b8d-38e36bc6d26c/invoke \"HTTP/1.1 200 OK\"\n", - "100%|████████████████████████████████████████| 7340/7340 [268:13<0:00, 27.4 steps/min]\n", - "INFO:httpx:HTTP Request: POST https://orchestration.hud.so/hud-gym/api/v2/environments/d71be89e-00e2-40e7-8b8d-38e36bc6d26c/close \"HTTP/1.1 200 OK\"\n", - "INFO:httpx:HTTP Request: GET https://orchestration.hud.so/hud-gym/api/v2/jobs/a2c1347a-2925-45ed-b86a-6b475b0dc4eb/trajectories \"HTTP/1.1 200 OK\"\n" - ] - }, - { - "name": "stdout", - "output_type": "stream", - "text": [ - "{'task_count': 360, 'avg_reward': 0.21677517254432735, 'success_rate': 18.333333333333332}\n", - "View results at: https://app.hud.so/jobs/a2c1347a-2925-45ed-b86a-6b475b0dc4eb\n" + "\u001b[90m╔══════════════════════════════════════════════════════════════════════════════════════════════════╗\u001b[0m\n", + "\u001b[90m║\u001b[0m 🚀 Job 'osworld b4db80 openai/computer-use-preview+anthropic/claude-opus-4-1-20250805' started: \u001b[90m║\u001b[0m\n", + "\u001b[90m╟──────────────────────────────────────────────────────────────────────────────────────────────────╢\u001b[0m\n", + "\u001b[90m║\u001b[0m \u001b[1m\u001b[33mhttps://app.hud.so/jobs/0d10d7d0-2c86-4a5d-a36b-2c28719773cf\u001b[0m \u001b[90m║\u001b[0m\n", + "\u001b[90m╚══════════════════════════════════════════════════════════════════════════════════════════════════╝\u001b[0m\n", + "\n" + ] + }, + { + "name": "stderr", + "output_type": "stream", + "text": [ + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Step failed: litellm.InternalServerError: InternalServerError: OpenAIException - {\n", + " \"error\": {\n", + " \"message\": \"An error occurred while processing your request. You can retry your request, or contact us through our help center at help.openai.com if the error persists. Please include the request ID req_97cc086a1b58a101f7db3ea88323a12f in your message.\",\n", + " \"type\": \"server_error\",\n", + " \"param\": null,\n", + " \"code\": \"server_error\"\n", + " }\n", + "}\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Step failed: 22 validation errors for ResponseComputerToolCall\n", + "action.ActionClick.button\n", + " Field required [type=missing, input_value={'keys': 'Tab', 'type': 'keypress'}, input_type=dict]\n", + " For further information visit https://errors.pydantic.dev/2.11/v/missing\n", + "action.ActionClick.type\n", + " Input should be 'click' [type=literal_error, input_value='keypress', input_type=str]\n", + " For further information visit https://errors.pydantic.dev/2.11/v/literal_error\n", + "action.ActionClick.x\n", + " Field required [type=missing, input_value={'keys': 'Tab', 'type': 'keypress'}, input_type=dict]\n", + " For further information visit https://errors.pydantic.dev/2.11/v/missing\n", + "action.ActionClick.y\n", + " Field required [type=missing, input_value={'keys': 'Tab', 'type': 'keypress'}, input_type=dict]\n", + " For further information visit https://errors.pydantic.dev/2.11/v/missing\n", + "action.ActionDoubleClick.type\n", + " Input should be 'double_click' [type=literal_error, input_value='keypress', input_type=str]\n", + " For further information visit https://errors.pydantic.dev/2.11/v/literal_error\n", + "action.ActionDoubleClick.x\n", + " Field required [type=missing, input_value={'keys': 'Tab', 'type': 'keypress'}, input_type=dict]\n", + " For further information visit https://errors.pydantic.dev/2.11/v/missing\n", + "action.ActionDoubleClick.y\n", + " Field required [type=missing, input_value={'keys': 'Tab', 'type': 'keypress'}, input_type=dict]\n", + " For further information visit https://errors.pydantic.dev/2.11/v/missing\n", + "action.ActionDrag.path\n", + " Field required [type=missing, input_value={'keys': 'Tab', 'type': 'keypress'}, input_type=dict]\n", + " For further information visit https://errors.pydantic.dev/2.11/v/missing\n", + "action.ActionDrag.type\n", + " Input should be 'drag' [type=literal_error, input_value='keypress', input_type=str]\n", + " For further information visit https://errors.pydantic.dev/2.11/v/literal_error\n", + "action.ActionKeypress.keys\n", + " Input should be a valid list [type=list_type, input_value='Tab', input_type=str]\n", + " For further information visit https://errors.pydantic.dev/2.11/v/list_type\n", + "action.ActionMove.type\n", + " Input should be 'move' [type=literal_error, input_value='keypress', input_type=str]\n", + " For further information visit https://errors.pydantic.dev/2.11/v/literal_error\n", + "action.ActionMove.x\n", + " Field required [type=missing, input_value={'keys': 'Tab', 'type': 'keypress'}, input_type=dict]\n", + " For further information visit https://errors.pydantic.dev/2.11/v/missing\n", + "action.ActionMove.y\n", + " Field required [type=missing, input_value={'keys': 'Tab', 'type': 'keypress'}, input_type=dict]\n", + " For further information visit https://errors.pydantic.dev/2.11/v/missing\n", + "action.ActionScreenshot.type\n", + " Input should be 'screenshot' [type=literal_error, input_value='keypress', input_type=str]\n", + " For further information visit https://errors.pydantic.dev/2.11/v/literal_error\n", + "action.ActionScroll.scroll_x\n", + " Field required [type=missing, input_value={'keys': 'Tab', 'type': 'keypress'}, input_type=dict]\n", + " For further information visit https://errors.pydantic.dev/2.11/v/missing\n", + "action.ActionScroll.scroll_y\n", + " Field required [type=missing, input_value={'keys': 'Tab', 'type': 'keypress'}, input_type=dict]\n", + " For further information visit https://errors.pydantic.dev/2.11/v/missing\n", + "action.ActionScroll.type\n", + " Input should be 'scroll' [type=literal_error, input_value='keypress', input_type=str]\n", + " For further information visit https://errors.pydantic.dev/2.11/v/literal_error\n", + "action.ActionScroll.x\n", + " Field required [type=missing, input_value={'keys': 'Tab', 'type': 'keypress'}, input_type=dict]\n", + " For further information visit https://errors.pydantic.dev/2.11/v/missing\n", + "action.ActionScroll.y\n", + " Field required [type=missing, input_value={'keys': 'Tab', 'type': 'keypress'}, input_type=dict]\n", + " For further information visit https://errors.pydantic.dev/2.11/v/missing\n", + "action.ActionType.text\n", + " Field required [type=missing, input_value={'keys': 'Tab', 'type': 'keypress'}, input_type=dict]\n", + " For further information visit https://errors.pydantic.dev/2.11/v/missing\n", + "action.ActionType.type\n", + " Input should be 'type' [type=literal_error, input_value='keypress', input_type=str]\n", + " For further information visit https://errors.pydantic.dev/2.11/v/literal_error\n", + "action.ActionWait.type\n", + " Input should be 'wait' [type=literal_error, input_value='keypress', input_type=str]\n", + " For further information visit https://errors.pydantic.dev/2.11/v/literal_error\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Step failed: litellm.InternalServerError: InternalServerError: OpenAIException - {\n", + " \"error\": {\n", + " \"message\": \"An error occurred while processing your request. You can retry your request, or contact us through our help center at help.openai.com if the error persists. Please include the request ID req_f7a393a984e7e85dc5845aef8a9471e4 in your message.\",\n", + " \"type\": \"model_error\",\n", + " \"param\": null,\n", + " \"code\": null\n", + " }\n", + "}\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Step failed: litellm.InternalServerError: AnthropicError - {\"type\":\"error\",\"error\":{\"type\":\"overloaded_error\",\"message\":\"Overloaded\"},\"request_id\":null}\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Client is not running, cannot disconnect\n", + "Client is not running, cannot disconnect\n", + "Client is not running, cannot disconnect\n", + "Client is not running, cannot disconnect\n", + "Client is not running, cannot disconnect\n", + "Client is not running, cannot disconnect\n", + "Client is not running, cannot disconnect\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n", + "Error parsing JSON response\n", + "Traceback (most recent call last):\n", + " File \"c:\\Users\\dillo\\miniconda3\\envs\\cua\\Lib\\site-packages\\mcp\\client\\streamable_http.py\", line 310, in _handle_json_response\n", + " await read_stream_writer.send(session_message)\n", + " File \"c:\\Users\\dillo\\miniconda3\\envs\\cua\\Lib\\site-packages\\anyio\\streams\\memory.py\", line 242, in send\n", + " self.send_nowait(item)\n", + " File \"c:\\Users\\dillo\\miniconda3\\envs\\cua\\Lib\\site-packages\\anyio\\streams\\memory.py\", line 211, in send_nowait\n", + " raise ClosedResourceError\n", + "anyio.ClosedResourceError\n", + "Request handler error: \n", + "Failed to close auto-created client: \n", + "Error parsing JSON response\n", + "Traceback (most recent call last):\n", + " File \"c:\\Users\\dillo\\miniconda3\\envs\\cua\\Lib\\site-packages\\mcp\\client\\streamable_http.py\", line 310, in _handle_json_response\n", + " await read_stream_writer.send(session_message)\n", + " File \"c:\\Users\\dillo\\miniconda3\\envs\\cua\\Lib\\site-packages\\anyio\\streams\\memory.py\", line 242, in send\n", + " self.send_nowait(item)\n", + " File \"c:\\Users\\dillo\\miniconda3\\envs\\cua\\Lib\\site-packages\\anyio\\streams\\memory.py\", line 211, in send_nowait\n", + " raise ClosedResourceError\n", + "anyio.ClosedResourceError\n", + "Request handler error: \n", + "Failed to close auto-created client: \n", + "Tool evaluate has an output schema but did not return structured content. Continuing without structured content validation.\n" + ] + }, + { + "ename": "CancelledError", + "evalue": "", + "output_type": "error", + "traceback": [ + "\u001b[1;31m---------------------------------------------------------------------------\u001b[0m", + "\u001b[1;31mCancelledError\u001b[0m Traceback (most recent call last)", + "Cell \u001b[1;32mIn[7], line 14\u001b[0m\n\u001b[0;32m 11\u001b[0m job_uuid \u001b[38;5;241m=\u001b[39m \u001b[38;5;28mstr\u001b[39m(uuid\u001b[38;5;241m.\u001b[39muuid4())[:\u001b[38;5;241m6\u001b[39m]\n\u001b[0;32m 12\u001b[0m job_name \u001b[38;5;241m=\u001b[39m \u001b[38;5;124mf\u001b[39m\u001b[38;5;124m\"\u001b[39m\u001b[38;5;124mosworld \u001b[39m\u001b[38;5;132;01m{\u001b[39;00mjob_uuid\u001b[38;5;132;01m}\u001b[39;00m\u001b[38;5;124m \u001b[39m\u001b[38;5;132;01m{\u001b[39;00mmodel\u001b[38;5;132;01m}\u001b[39;00m\u001b[38;5;124m\"\u001b[39m\n\u001b[1;32m---> 14\u001b[0m results \u001b[38;5;241m=\u001b[39m \u001b[38;5;28;01mawait\u001b[39;00m run_full_dataset(\n\u001b[0;32m 15\u001b[0m dataset\u001b[38;5;241m=\u001b[39m\u001b[38;5;124m\"\u001b[39m\u001b[38;5;124mhud-evals/OSWorld-Verified-XLang\u001b[39m\u001b[38;5;124m\"\u001b[39m,\n\u001b[0;32m 16\u001b[0m job_name\u001b[38;5;241m=\u001b[39mjob_name, \n\u001b[0;32m 17\u001b[0m model\u001b[38;5;241m=\u001b[39mmodel,\n\u001b[0;32m 18\u001b[0m max_concurrent\u001b[38;5;241m=\u001b[39m\u001b[38;5;241m30\u001b[39m, \n\u001b[0;32m 19\u001b[0m max_steps\u001b[38;5;241m=\u001b[39m\u001b[38;5;241m75\u001b[39m,\n\u001b[0;32m 20\u001b[0m trajectory_dir\u001b[38;5;241m=\u001b[39m\u001b[38;5;124mf\u001b[39m\u001b[38;5;124m\"\u001b[39m\u001b[38;5;124mtrajectories/osworld_\u001b[39m\u001b[38;5;132;01m{\u001b[39;00mjob_uuid\u001b[38;5;132;01m}\u001b[39;00m\u001b[38;5;124m\"\u001b[39m,\n\u001b[0;32m 21\u001b[0m )\n\u001b[0;32m 23\u001b[0m \u001b[38;5;66;03m# results is a list from hud.datasets.run_dataset; inspect/aggregate as needed\u001b[39;00m\n\u001b[0;32m 24\u001b[0m \u001b[38;5;28mprint\u001b[39m(\u001b[38;5;124mf\u001b[39m\u001b[38;5;124m\"\u001b[39m\u001b[38;5;124mJob: \u001b[39m\u001b[38;5;132;01m{\u001b[39;00mjob_name\u001b[38;5;132;01m}\u001b[39;00m\u001b[38;5;124m\"\u001b[39m)\n", + "File \u001b[1;32mF:\\Projects\\cua\\cua-clean\\libs\\python\\agent\\agent\\integrations\\hud\\__init__.py:134\u001b[0m, in \u001b[0;36mrun_full_dataset\u001b[1;34m(dataset, job_name, model, allowed_tools, max_concurrent, max_steps, split, trajectory_dir)\u001b[0m\n\u001b[0;32m 131\u001b[0m job_name \u001b[38;5;241m=\u001b[39m job_name \u001b[38;5;129;01mor\u001b[39;00m \u001b[38;5;124mf\u001b[39m\u001b[38;5;124m\"\u001b[39m\u001b[38;5;124mEvaluation \u001b[39m\u001b[38;5;132;01m{\u001b[39;00mtime\u001b[38;5;241m.\u001b[39mstrftime(\u001b[38;5;124m'\u001b[39m\u001b[38;5;124m%\u001b[39m\u001b[38;5;124mH:\u001b[39m\u001b[38;5;124m%\u001b[39m\u001b[38;5;124mM \u001b[39m\u001b[38;5;124m%\u001b[39m\u001b[38;5;124mY-\u001b[39m\u001b[38;5;124m%\u001b[39m\u001b[38;5;124mm-\u001b[39m\u001b[38;5;132;01m%d\u001b[39;00m\u001b[38;5;124m'\u001b[39m)\u001b[38;5;132;01m}\u001b[39;00m\u001b[38;5;124m\"\u001b[39m\n\u001b[0;32m 133\u001b[0m \u001b[38;5;66;03m# Execute evaluation\u001b[39;00m\n\u001b[1;32m--> 134\u001b[0m \u001b[38;5;28;01mreturn\u001b[39;00m \u001b[38;5;28;01mawait\u001b[39;00m run_dataset(\n\u001b[0;32m 135\u001b[0m name\u001b[38;5;241m=\u001b[39mjob_name,\n\u001b[0;32m 136\u001b[0m dataset\u001b[38;5;241m=\u001b[39mdataset,\n\u001b[0;32m 137\u001b[0m agent_class\u001b[38;5;241m=\u001b[39mProxyOperatorAgent,\n\u001b[0;32m 138\u001b[0m agent_config\u001b[38;5;241m=\u001b[39m{\u001b[38;5;124m\"\u001b[39m\u001b[38;5;124mmodel\u001b[39m\u001b[38;5;124m\"\u001b[39m: model, \u001b[38;5;124m\"\u001b[39m\u001b[38;5;124mallowed_tools\u001b[39m\u001b[38;5;124m\"\u001b[39m: allowed_tools, \u001b[38;5;124m\"\u001b[39m\u001b[38;5;124mtrajectory_dir\u001b[39m\u001b[38;5;124m\"\u001b[39m: trajectory_dir},\n\u001b[0;32m 139\u001b[0m max_concurrent\u001b[38;5;241m=\u001b[39mmax_concurrent,\n\u001b[0;32m 140\u001b[0m metadata\u001b[38;5;241m=\u001b[39m{\u001b[38;5;124m\"\u001b[39m\u001b[38;5;124mdataset\u001b[39m\u001b[38;5;124m\"\u001b[39m: dataset_name},\n\u001b[0;32m 141\u001b[0m max_steps\u001b[38;5;241m=\u001b[39mmax_steps,\n\u001b[0;32m 142\u001b[0m auto_respond\u001b[38;5;241m=\u001b[39m\u001b[38;5;28;01mTrue\u001b[39;00m,\n\u001b[0;32m 143\u001b[0m )\n", + "File \u001b[1;32mc:\\Users\\dillo\\miniconda3\\envs\\cua\\Lib\\site-packages\\hud\\datasets.py:262\u001b[0m, in \u001b[0;36mrun_dataset\u001b[1;34m(name, dataset, agent_class, agent_config, max_concurrent, metadata, max_steps, split, auto_respond, custom_system_prompt)\u001b[0m\n\u001b[0;32m 259\u001b[0m results[index] \u001b[38;5;241m=\u001b[39m \u001b[38;5;28;01mawait\u001b[39;00m agent\u001b[38;5;241m.\u001b[39mrun(task, max_steps\u001b[38;5;241m=\u001b[39mmax_steps)\n\u001b[0;32m 261\u001b[0m \u001b[38;5;66;03m# Execute all tasks\u001b[39;00m\n\u001b[1;32m--> 262\u001b[0m \u001b[38;5;28;01mawait\u001b[39;00m asyncio\u001b[38;5;241m.\u001b[39mgather(\n\u001b[0;32m 263\u001b[0m \u001b[38;5;241m*\u001b[39m[_worker(i, task, max_steps\u001b[38;5;241m=\u001b[39mmax_steps) \u001b[38;5;28;01mfor\u001b[39;00m i, task \u001b[38;5;129;01min\u001b[39;00m \u001b[38;5;28menumerate\u001b[39m(dataset)],\n\u001b[0;32m 264\u001b[0m return_exceptions\u001b[38;5;241m=\u001b[39m\u001b[38;5;28;01mTrue\u001b[39;00m, \u001b[38;5;66;03m# Don't fail entire batch on one error\u001b[39;00m\n\u001b[0;32m 265\u001b[0m )\n\u001b[0;32m 267\u001b[0m \u001b[38;5;28;01mreturn\u001b[39;00m results\n", + "File \u001b[1;32mc:\\Users\\dillo\\miniconda3\\envs\\cua\\Lib\\site-packages\\hud\\datasets.py:246\u001b[0m, in \u001b[0;36mrun_dataset.._worker\u001b[1;34m(index, task_dict, max_steps)\u001b[0m\n\u001b[0;32m 245\u001b[0m \u001b[38;5;28;01masync\u001b[39;00m \u001b[38;5;28;01mdef\u001b[39;00m\u001b[38;5;250m \u001b[39m\u001b[38;5;21m_worker\u001b[39m(index: \u001b[38;5;28mint\u001b[39m, task_dict: Any, max_steps: \u001b[38;5;28mint\u001b[39m \u001b[38;5;241m=\u001b[39m \u001b[38;5;241m40\u001b[39m) \u001b[38;5;241m-\u001b[39m\u001b[38;5;241m>\u001b[39m \u001b[38;5;28;01mNone\u001b[39;00m:\n\u001b[1;32m--> 246\u001b[0m \u001b[38;5;28;01masync\u001b[39;00m \u001b[38;5;28;01mwith\u001b[39;00m sem:\n\u001b[0;32m 247\u001b[0m \u001b[38;5;66;03m# Create trace for this task\u001b[39;00m\n\u001b[0;32m 248\u001b[0m task_name \u001b[38;5;241m=\u001b[39m task_dict\u001b[38;5;241m.\u001b[39mget(\u001b[38;5;124m\"\u001b[39m\u001b[38;5;124mprompt\u001b[39m\u001b[38;5;124m\"\u001b[39m) \u001b[38;5;129;01mor\u001b[39;00m \u001b[38;5;124mf\u001b[39m\u001b[38;5;124m\"\u001b[39m\u001b[38;5;124mTask \u001b[39m\u001b[38;5;132;01m{\u001b[39;00mindex\u001b[38;5;132;01m}\u001b[39;00m\u001b[38;5;124m\"\u001b[39m\n\u001b[0;32m 249\u001b[0m \u001b[38;5;28;01mif\u001b[39;00m \u001b[38;5;124m\"\u001b[39m\u001b[38;5;124msystem_prompt\u001b[39m\u001b[38;5;124m\"\u001b[39m \u001b[38;5;129;01mnot\u001b[39;00m \u001b[38;5;129;01min\u001b[39;00m task_dict:\n", + "File \u001b[1;32mc:\\Users\\dillo\\miniconda3\\envs\\cua\\Lib\\asyncio\\locks.py:14\u001b[0m, in \u001b[0;36m_ContextManagerMixin.__aenter__\u001b[1;34m(self)\u001b[0m\n\u001b[0;32m 13\u001b[0m \u001b[38;5;28;01masync\u001b[39;00m \u001b[38;5;28;01mdef\u001b[39;00m\u001b[38;5;250m \u001b[39m\u001b[38;5;21m__aenter__\u001b[39m(\u001b[38;5;28mself\u001b[39m):\n\u001b[1;32m---> 14\u001b[0m \u001b[38;5;28;01mawait\u001b[39;00m \u001b[38;5;28mself\u001b[39m\u001b[38;5;241m.\u001b[39macquire()\n\u001b[0;32m 15\u001b[0m \u001b[38;5;66;03m# We have no use for the \"as ...\" clause in the with\u001b[39;00m\n\u001b[0;32m 16\u001b[0m \u001b[38;5;66;03m# statement for locks.\u001b[39;00m\n\u001b[0;32m 17\u001b[0m \u001b[38;5;28;01mreturn\u001b[39;00m \u001b[38;5;28;01mNone\u001b[39;00m\n", + "File \u001b[1;32mc:\\Users\\dillo\\miniconda3\\envs\\cua\\Lib\\asyncio\\locks.py:386\u001b[0m, in \u001b[0;36mSemaphore.acquire\u001b[1;34m(self)\u001b[0m\n\u001b[0;32m 384\u001b[0m \u001b[38;5;28;01mtry\u001b[39;00m:\n\u001b[0;32m 385\u001b[0m \u001b[38;5;28;01mtry\u001b[39;00m:\n\u001b[1;32m--> 386\u001b[0m \u001b[38;5;28;01mawait\u001b[39;00m fut\n\u001b[0;32m 387\u001b[0m \u001b[38;5;28;01mfinally\u001b[39;00m:\n\u001b[0;32m 388\u001b[0m \u001b[38;5;28mself\u001b[39m\u001b[38;5;241m.\u001b[39m_waiters\u001b[38;5;241m.\u001b[39mremove(fut)\n", + "\u001b[1;31mCancelledError\u001b[0m: " ] } ], "source": [ - "from agent.integrations.hud import run_job\n", - "from hud import load_taskset\n", - "from hud.taskset import TaskSet\n", - "import logging\n", "import uuid\n", + "from agent.integrations.hud import run_full_dataset\n", "\n", - "# Load taskset\n", - "taskset = await load_taskset(\"OSWorld-Verified\")\n", - "# taskset = TaskSet(tasks=taskset[:20]) # limit to 10 tasks instead of all 370\n", + "models_to_test = [\n", + " \"openai/computer-use-preview+anthropic/claude-opus-4-20250514\",\n", + "]\n", + " \n", "\n", - "job_name = \"osworld-gta-gpt5\"\n", - "job_name = f\"{job_name}-{str(uuid.uuid4())[:4]}\"\n", + "for model in models_to_test:\n", + " # Full dataset evaluation (runs via HUD's run_dataset under the hood)\n", + " job_uuid = str(uuid.uuid4())[:6]\n", + " job_name = f\"osworld {job_uuid} {model}\"\n", "\n", - "# Run benchmark job\n", - "job = await run_job(\n", - " # model=\"openai/computer-use-preview\",\n", - " model=\"huggingface-local/HelloKKMe/GTA1-7B+openai/gpt-5\",\n", - " task_or_taskset=taskset,\n", - " job_name=job_name,\n", - " max_concurrent_tasks=20,\n", - " # add any extra ComputerAgent kwargs:\n", - " verbosity=logging.INFO, # Enable logging\n", - " trajectory_dir=f\"trajectories/{job_name}\" # Save trajectories locally\n", - ")\n", + " results = await run_full_dataset(\n", + " dataset=\"hud-evals/OSWorld-Verified-XLang\",\n", + " job_name=job_name, \n", + " model=model,\n", + " max_concurrent=20, \n", + " max_steps=75,\n", + " trajectory_dir=f\"trajectories/osworld_{job_uuid}\",\n", + " only_n_most_recent_images=3\n", + " )\n", "\n", - "# Get results OR view them at app.hud.so\n", - "print(await job.get_analytics())\n", - "print(f\"View results at: https://app.hud.so/jobs/{job.id}\")" + " # results is a list from hud.datasets.run_dataset; inspect/aggregate as needed\n", + " print(f\"Job: {job_name}\")\n", + " print(f\"Total results: {len(results)}\")" ] } ],