Basenull AI Ops · Multi-Agent Infrastructure

Make sure your AI agents actually finish the job.

Process monitoring for AI agents. Define a workflow, watch your agent run it, get alerted when it skips a step or stops running on time.

Every team is shipping agents. Most pay attention the first few weeks, then move on. Months later the model drifts, an upstream API changes, and your workflow is silently broken — until something downstream visibly fails. agent-check catches the drift the day it happens, not three weeks later.

Free · 1 workflow · 100 verifier credits · no credit card

How it works

Define a workflow

Paste your agent's existing instruction prompt; we draft a list of tasks with expected evidence for each. Tweak, save.

Wire your agent

Drop the MCP config snippet into your agent's setup. The agent calls start_run, submits evidence per task, calls complete_run.

Get alerted when it breaks

A small LLM verifies each piece of evidence against your description. Failure → email, Slack, or webhook within minutes. Matrix view shows the cadence.

What gets monitored

Per-task evidence verification

Each task has an expected_evidence description. The agent submits text evidence; a small LLM verifies it matches what you described. Optional: a local verify_script runs through the MCP and submits its stdout instead.

Cadence detection

Workflow scheduled daily/weekly/cron? We notice when a run fails to start by its deadline + grace window. No more 'wait, has that agent run this month?'

Drift alerts in real time

Email, Slack incoming-webhook, or generic webhook. Configure per workflow per alert type. Hear about a failed verification within minutes, not weeks.

SOC 2-grade audit trail

Every run, every submitted evidence, every verifier verdict — preserved per your retention tier. CSV + JSON export. Reviewable trail when compliance asks.

Workflow matrix view

Tasks across columns, runs down rows, status icons per cell. Five seconds from open to 'I see something is wrong.' Drill into any cell to see the actual evidence.

Embeddable status badge

Drop the SVG badge in your README. Every README view = a passing/failing pulse on your agent's health, visible to your team and the world.

Your agent claimed "task complete".

Was it? agent-check is the layer above per-call observability — workflow process monitoring, MCP-native, with AI-validated evidence per task.

Sign in to create a workflow