Files
hatchet/examples/python/bulk_operations/test_bulk_replay.py
Matt Kaye a73e34cd92 Feat: Batched replays (#1860)
* feat: batched replays

* fix: add some comments

* feat: e2e test for bulk replays

* chore: gen

* fix: improving tests a bit

* fix: copilot

Co-authored-by: Copilot <175728472+Copilot@users.noreply.github.com>

* chore: gen

* fix: tag

---------

Co-authored-by: Copilot <175728472+Copilot@users.noreply.github.com>
2025-06-13 13:47:50 -04:00

107 lines
3.1 KiB
Python

import asyncio
from datetime import datetime, timedelta, timezone
from uuid import uuid4
import pytest
from examples.bulk_operations.worker import (
bulk_replay_test_1,
bulk_replay_test_2,
bulk_replay_test_3,
)
from hatchet_sdk import BulkCancelReplayOpts, Hatchet, RunFilter, TriggerWorkflowOptions
from hatchet_sdk.clients.rest.models.v1_task_status import V1TaskStatus
@pytest.mark.asyncio(loop_scope="session")
async def test_bulk_replay(hatchet: Hatchet) -> None:
test_run_id = str(uuid4())
n = 100
with pytest.raises(Exception):
await bulk_replay_test_1.aio_run_many(
[
bulk_replay_test_1.create_bulk_run_item(
options=TriggerWorkflowOptions(
additional_metadata={
"test_run_id": test_run_id,
}
)
)
for _ in range(n + 1)
]
)
with pytest.raises(Exception):
await bulk_replay_test_2.aio_run_many(
[
bulk_replay_test_2.create_bulk_run_item(
options=TriggerWorkflowOptions(
additional_metadata={
"test_run_id": test_run_id,
}
)
)
for _ in range((n // 2) - 1)
]
)
with pytest.raises(Exception):
await bulk_replay_test_3.aio_run_many(
[
bulk_replay_test_3.create_bulk_run_item(
options=TriggerWorkflowOptions(
additional_metadata={
"test_run_id": test_run_id,
}
)
)
for _ in range((n // 2) - 2)
]
)
workflow_ids = [
bulk_replay_test_1.id,
bulk_replay_test_2.id,
bulk_replay_test_3.id,
]
## Should result in two batches of replays
await hatchet.runs.aio_bulk_replay(
opts=BulkCancelReplayOpts(
filters=RunFilter(
workflow_ids=workflow_ids,
since=datetime.now(tz=timezone.utc) - timedelta(minutes=2),
additional_metadata={"test_run_id": test_run_id},
)
)
)
await asyncio.sleep(5)
runs = await hatchet.runs.aio_list(
workflow_ids=workflow_ids,
since=datetime.now(tz=timezone.utc) - timedelta(minutes=2),
additional_metadata={"test_run_id": test_run_id},
limit=1000,
)
assert len(runs.rows) == n + 1 + (n // 2 - 1) + (n // 2 - 2)
for run in runs.rows:
assert run.status == V1TaskStatus.COMPLETED
assert run.retry_count == 1
assert run.attempt == 2
assert (
len([r for r in runs.rows if r.workflow_id == bulk_replay_test_1.id]) == n + 1
)
assert (
len([r for r in runs.rows if r.workflow_id == bulk_replay_test_2.id])
== n // 2 - 1
)
assert (
len([r for r in runs.rows if r.workflow_id == bulk_replay_test_3.id])
== n // 2 - 2
)