fix: feat: implement evidence/resources and evidence/protocol logging (#1059)
- Add evidence/resources/ and evidence/protocol/ directories with .gitkeep - Add schemas for resources/ and protocol/ to evidence/README.md - Create formulas/run-resources.toml (sense formula: disk/RAM/API/CI metrics, daily cron 06:00 UTC, verdict: ok/warn/critical) - Create formulas/run-protocol.toml (sense formula: TVL/fees/positions/ rebalance frequency via LmTotalEth.s.sol + cast, daily cron 07:00 UTC, verdict: healthy/degraded/offline) - Update STATE.md Co-Authored-By: Claude Sonnet 4.6 <noreply@anthropic.com>
This commit is contained in:
parent
7fd1963480
commit
de014e9b13
6 changed files with 458 additions and 1 deletions
187
formulas/run-protocol.toml
Normal file
187
formulas/run-protocol.toml
Normal file
|
|
@ -0,0 +1,187 @@
|
|||
# formulas/run-protocol.toml
|
||||
#
|
||||
# On-chain protocol health snapshot — collect TVL, accumulated fees,
|
||||
# position count, and rebalance frequency from the deployed LiquidityManager.
|
||||
# Write a structured JSON evidence file for planner and predictor consumption.
|
||||
#
|
||||
# Type: sense. Read-only — produces metrics only, no git artifacts.
|
||||
#
|
||||
# Staleness threshold: 1 day (matches evidence/protocol/ schema).
|
||||
# Cron: daily at 07:00 UTC (staggered 1 h after run-resources).
|
||||
|
||||
[formula]
|
||||
id = "run-protocol"
|
||||
name = "On-Chain Protocol Health Snapshot"
|
||||
description = "Collect TVL, accumulated fees, position count, and rebalance frequency from the deployed LiquidityManager; write evidence/protocol/{date}.json."
|
||||
type = "sense"
|
||||
# "sense" → read-only, produces metrics only
|
||||
# "act" → produces git artifacts (cf. run-evolution, run-red-team)
|
||||
|
||||
# ── Cron ───────────────────────────────────────────────────────────────────────
|
||||
|
||||
[cron]
|
||||
schedule = "0 7 * * *" # daily at 07:00 UTC (1 h after run-resources)
|
||||
description = "Matches 1-day staleness threshold — one snapshot per day keeps the record fresh."
|
||||
|
||||
# ── Inputs ─────────────────────────────────────────────────────────────────────
|
||||
|
||||
[inputs.rpc_url]
|
||||
type = "string"
|
||||
required = true
|
||||
description = """
|
||||
Base network RPC endpoint used to query on-chain state.
|
||||
Example: https://mainnet.base.org or a running Anvil fork URL.
|
||||
"""
|
||||
|
||||
[inputs.deployments_file]
|
||||
type = "string"
|
||||
required = false
|
||||
default = "onchain/deployments-local.json"
|
||||
description = """
|
||||
Path to the deployments JSON file containing contract addresses.
|
||||
The formula reads LiquidityManager address from this file.
|
||||
Use onchain/deployments.json for mainnet; onchain/deployments-local.json
|
||||
for a local Anvil fork.
|
||||
"""
|
||||
|
||||
[inputs.lookback_blocks]
|
||||
type = "integer"
|
||||
required = false
|
||||
default = 7200
|
||||
description = """
|
||||
Number of blocks to scan for Recenter events when computing
|
||||
rebalance_count_24h (~24 h of Base blocks at ~2 s/block).
|
||||
"""
|
||||
|
||||
# ── Execution ──────────────────────────────────────────────────────────────────
|
||||
|
||||
[execution]
|
||||
script = "scripts/harb-evaluator/run-protocol.sh"
|
||||
invocation = "RPC_URL={rpc_url} DEPLOYMENTS_FILE={deployments_file} LOOKBACK_BLOCKS={lookback_blocks} bash scripts/harb-evaluator/run-protocol.sh"
|
||||
|
||||
# Exit codes:
|
||||
# 0 snapshot written successfully
|
||||
# 2 infrastructure error (RPC unreachable, missing deployments file, forge unavailable, etc.)
|
||||
|
||||
# ── Steps ──────────────────────────────────────────────────────────────────────
|
||||
|
||||
[[steps]]
|
||||
id = "read-addresses"
|
||||
description = """
|
||||
Read the LiquidityManager contract address from {deployments_file}.
|
||||
Fail with exit code 2 if the file is absent or the address is missing.
|
||||
"""
|
||||
|
||||
[[steps]]
|
||||
id = "collect-tvl"
|
||||
description = """
|
||||
Query LiquidityManager total ETH via forge script LmTotalEth.s.sol
|
||||
against {rpc_url}.
|
||||
Records tvl_eth (wei string) and tvl_eth_formatted (ETH, 2 dp).
|
||||
LmTotalEth.s.sol uses exact Uniswap V3 integer math (LiquidityAmounts +
|
||||
TickMath) to sum free ETH, free WETH, and ETH locked across all three
|
||||
positions (floor, anchor, discovery).
|
||||
"""
|
||||
forge_script = "onchain/script/LmTotalEth.s.sol"
|
||||
|
||||
[[steps]]
|
||||
id = "collect-fees"
|
||||
description = """
|
||||
Query accumulated protocol fees from the LiquidityManager via cast call:
|
||||
LiquidityManager.accumulatedFees() → uint256
|
||||
Records accumulated_fees_eth (wei string) and accumulated_fees_eth_formatted
|
||||
(ETH, 3 dp).
|
||||
Falls back to 0 gracefully if the function is not present on the deployed
|
||||
contract (older deployment without fee tracking).
|
||||
"""
|
||||
|
||||
[[steps]]
|
||||
id = "collect-positions"
|
||||
description = """
|
||||
Query the three Uniswap V3 positions held by the LiquidityManager:
|
||||
LiquidityManager.floorPosition() → (tickLower, tickUpper, liquidity)
|
||||
LiquidityManager.anchorPosition() → (tickLower, tickUpper, liquidity)
|
||||
LiquidityManager.discoveryPosition() → (tickLower, tickUpper, liquidity)
|
||||
Records position_count (number of positions with liquidity > 0) and the
|
||||
positions array.
|
||||
"""
|
||||
|
||||
[[steps]]
|
||||
id = "collect-rebalances"
|
||||
description = """
|
||||
Count Recenter events emitted by the LiquidityManager in the past
|
||||
{lookback_blocks} blocks via eth_getLogs.
|
||||
Records:
|
||||
- rebalance_count_24h: total Recenter event count in the window.
|
||||
- last_rebalance_block: block number of the most recent Recenter event
|
||||
(0 if none found in the window).
|
||||
"""
|
||||
event_signature = "Recenter(int24,int24,int24)"
|
||||
|
||||
[[steps]]
|
||||
id = "collect"
|
||||
description = """
|
||||
Assemble all collected metrics into evidence/protocol/{date}.json.
|
||||
Compute verdict:
|
||||
- "offline" if tvl_eth = 0 or RPC was unreachable.
|
||||
- "degraded" if position_count < 3, or rebalance_count_24h = 0 and the
|
||||
protocol has been live for > 1 day.
|
||||
- "healthy" otherwise.
|
||||
Write the file conforming to the schema in evidence/README.md
|
||||
## Schema: protocol/YYYY-MM-DD.json.
|
||||
"""
|
||||
output = "evidence/protocol/{date}.json"
|
||||
schema = "evidence/README.md" # see ## Schema: protocol/YYYY-MM-DD.json
|
||||
|
||||
[[steps]]
|
||||
id = "deliver"
|
||||
description = """
|
||||
Commit evidence/protocol/{date}.json to main.
|
||||
Post a one-line summary comment to the originating issue (if any):
|
||||
verdict, tvl_eth_formatted, accumulated_fees_eth_formatted,
|
||||
position_count, rebalance_count_24h.
|
||||
On "degraded" or "offline": highlight the failing dimension and its value.
|
||||
"""
|
||||
|
||||
# ── Products ───────────────────────────────────────────────────────────────────
|
||||
|
||||
[products.evidence_file]
|
||||
path = "evidence/protocol/{date}.json"
|
||||
delivery = "commit to main"
|
||||
schema = "evidence/README.md" # see ## Schema: protocol/YYYY-MM-DD.json
|
||||
|
||||
[products.issue_comment]
|
||||
delivery = "post to originating issue (if any)"
|
||||
content = "verdict, tvl_eth_formatted, accumulated_fees_eth_formatted, position_count, rebalance_count_24h"
|
||||
on_degraded = "highlight failing dimension and its current value"
|
||||
|
||||
# ── Resources ──────────────────────────────────────────────────────────────────
|
||||
|
||||
[resources]
|
||||
profile = "light"
|
||||
compute = "local — forge script + cast calls only; no Anvil or Docker startup required"
|
||||
rpc = "Base network RPC ({rpc_url}) — read-only calls"
|
||||
concurrency = "safe to run in parallel with other formulas"
|
||||
|
||||
# ── Notes ──────────────────────────────────────────────────────────────────────
|
||||
|
||||
[notes]
|
||||
tvl_metric = """
|
||||
TVL is measured as LiquidityManager total ETH: free ETH + free WETH + ETH
|
||||
locked across all three Uniswap V3 positions (floor, anchor, discovery).
|
||||
Uses the same LmTotalEth.s.sol forge script as run-red-team to ensure
|
||||
consistent measurement methodology.
|
||||
"""
|
||||
|
||||
rebalance_staleness = """
|
||||
A zero rebalance_count_24h on an established deployment indicates the
|
||||
recenter() upkeep bot (services/txnBot) has stalled. The "degraded"
|
||||
verdict triggers a planner alert. On a fresh deployment (< 1 day old)
|
||||
zero rebalances is expected and does not trigger degraded.
|
||||
"""
|
||||
|
||||
fees_fallback = """
|
||||
accumulated_fees_eth falls back to 0 for deployments without fee tracking.
|
||||
The verdict is not affected by a zero fee value alone — only TVL and
|
||||
position_count drive the verdict.
|
||||
"""
|
||||
155
formulas/run-resources.toml
Normal file
155
formulas/run-resources.toml
Normal file
|
|
@ -0,0 +1,155 @@
|
|||
# formulas/run-resources.toml
|
||||
#
|
||||
# Infrastructure resource snapshot — collect disk usage, RAM trends,
|
||||
# Anthropic API call counts and budget burn, and Woodpecker CI queue depth.
|
||||
# Write a structured JSON evidence file for planner and predictor consumption.
|
||||
#
|
||||
# Type: sense. Read-only — produces metrics only, no git artifacts.
|
||||
#
|
||||
# Staleness threshold: 1 day (matches evidence/resources/ schema).
|
||||
# Cron: daily at 06:00 UTC.
|
||||
|
||||
[formula]
|
||||
id = "run-resources"
|
||||
name = "Infrastructure Resource Snapshot"
|
||||
description = "Collect disk, RAM, API usage, Anthropic budget burn, and CI queue depth; write evidence/resources/{date}.json."
|
||||
type = "sense"
|
||||
# "sense" → read-only, produces metrics only
|
||||
# "act" → produces git artifacts (cf. run-evolution, run-red-team)
|
||||
|
||||
# ── Cron ───────────────────────────────────────────────────────────────────────
|
||||
|
||||
[cron]
|
||||
schedule = "0 6 * * *" # daily at 06:00 UTC
|
||||
description = "Matches 1-day staleness threshold — one snapshot per day keeps the record fresh."
|
||||
|
||||
# ── Inputs ─────────────────────────────────────────────────────────────────────
|
||||
|
||||
[inputs.disk_path]
|
||||
type = "string"
|
||||
required = false
|
||||
default = "/"
|
||||
description = "Filesystem path to measure disk usage for (passed to df)."
|
||||
|
||||
[inputs.anthropic_budget_usd_limit]
|
||||
type = "number"
|
||||
required = false
|
||||
default = 50.0
|
||||
description = "Configured Anthropic budget ceiling in USD. Used to compute budget_pct in the evidence record."
|
||||
|
||||
[inputs.woodpecker_api_url]
|
||||
type = "string"
|
||||
required = false
|
||||
default = "http://localhost:8090"
|
||||
description = "Base URL of the Woodpecker CI API. Set to empty string to skip CI metrics."
|
||||
|
||||
# ── Execution ──────────────────────────────────────────────────────────────────
|
||||
|
||||
[execution]
|
||||
script = "scripts/harb-evaluator/run-resources.sh"
|
||||
invocation = "DISK_PATH={disk_path} ANTHROPIC_BUDGET_USD_LIMIT={anthropic_budget_usd_limit} WOODPECKER_API_URL={woodpecker_api_url} bash scripts/harb-evaluator/run-resources.sh"
|
||||
|
||||
# Exit codes:
|
||||
# 0 snapshot written successfully
|
||||
# 2 infrastructure error (disk command unavailable, JSON write failed, etc.)
|
||||
|
||||
# ── Steps ──────────────────────────────────────────────────────────────────────
|
||||
|
||||
[[steps]]
|
||||
id = "collect-disk"
|
||||
description = """
|
||||
Measure disk usage on {disk_path} via `df -B1 {disk_path}`.
|
||||
Extract used_bytes, total_bytes, and used_pct.
|
||||
"""
|
||||
|
||||
[[steps]]
|
||||
id = "collect-ram"
|
||||
description = """
|
||||
Measure RAM usage via `free -b` (Linux) or `vm_stat` (macOS).
|
||||
Extract used_bytes, total_bytes, and used_pct.
|
||||
"""
|
||||
|
||||
[[steps]]
|
||||
id = "collect-api"
|
||||
description = """
|
||||
Collect Anthropic API metrics:
|
||||
- anthropic_calls_24h: count of API calls in the past 24 hours (read from
|
||||
tmp/anthropic-call-log.jsonl if present; 0 if absent).
|
||||
- anthropic_budget_usd_used: sum of cost_usd entries in the call log for
|
||||
the current calendar day (UTC); 0 if log absent.
|
||||
- anthropic_budget_usd_limit: from {anthropic_budget_usd_limit} input.
|
||||
- anthropic_budget_pct: used / limit * 100 (0 if limit = 0).
|
||||
"""
|
||||
call_log = "tmp/anthropic-call-log.jsonl"
|
||||
|
||||
[[steps]]
|
||||
id = "collect-ci"
|
||||
description = """
|
||||
Query Woodpecker CI API for queue state.
|
||||
GET {woodpecker_api_url}/api/queue/info:
|
||||
- woodpecker_queue_depth: length of the waiting queue.
|
||||
- woodpecker_running: count of currently running jobs.
|
||||
Skipped gracefully (fields set to null) when {woodpecker_api_url} is empty
|
||||
or the endpoint is unreachable.
|
||||
"""
|
||||
|
||||
[[steps]]
|
||||
id = "collect"
|
||||
description = """
|
||||
Assemble all collected metrics into evidence/resources/{date}.json.
|
||||
Compute verdict:
|
||||
- "critical" if disk_used_pct ≥ 95, ram_used_pct ≥ 95,
|
||||
or anthropic_budget_pct ≥ 95.
|
||||
- "warn" if disk_used_pct ≥ 80, ram_used_pct ≥ 80,
|
||||
or anthropic_budget_pct ≥ 80.
|
||||
- "ok" otherwise.
|
||||
Write the file conforming to the schema in evidence/README.md
|
||||
## Schema: resources/YYYY-MM-DD.json.
|
||||
"""
|
||||
output = "evidence/resources/{date}.json"
|
||||
schema = "evidence/README.md" # see ## Schema: resources/YYYY-MM-DD.json
|
||||
|
||||
[[steps]]
|
||||
id = "deliver"
|
||||
description = """
|
||||
Commit evidence/resources/{date}.json to main.
|
||||
Post a one-line summary comment to the originating issue (if any):
|
||||
verdict, disk_used_pct, ram_used_pct, anthropic_budget_pct, ci queue depth.
|
||||
On "warn" or "critical": highlight the breaching dimensions.
|
||||
"""
|
||||
|
||||
# ── Products ───────────────────────────────────────────────────────────────────
|
||||
|
||||
[products.evidence_file]
|
||||
path = "evidence/resources/{date}.json"
|
||||
delivery = "commit to main"
|
||||
schema = "evidence/README.md" # see ## Schema: resources/YYYY-MM-DD.json
|
||||
|
||||
[products.issue_comment]
|
||||
delivery = "post to originating issue (if any)"
|
||||
content = "verdict, disk_used_pct, ram_used_pct, anthropic_budget_pct, ci queue depth"
|
||||
on_warn = "highlight breaching dimensions and current values"
|
||||
|
||||
# ── Resources ──────────────────────────────────────────────────────────────────
|
||||
|
||||
[resources]
|
||||
profile = "light"
|
||||
compute = "local — shell commands only (df, free, curl); no Docker or Anvil required"
|
||||
concurrency = "safe to run in parallel with other formulas"
|
||||
|
||||
# ── Notes ──────────────────────────────────────────────────────────────────────
|
||||
|
||||
[notes]
|
||||
call_log = """
|
||||
tmp/anthropic-call-log.jsonl is expected to have one JSON object per line,
|
||||
each with at minimum:
|
||||
{ "ts": "<ISO timestamp>", "cost_usd": <number> }
|
||||
The file is written by the dark-factory agent loop. When absent the API
|
||||
metrics default to 0 — the snapshot is still written rather than failing.
|
||||
"""
|
||||
|
||||
disk_warn = """
|
||||
Planner MEMORY.md (2026-03-20) notes disk at 79%. The "warn" threshold
|
||||
(≥80%) will fire on the first run-resources pass. Monitor trajectory;
|
||||
evidence pipeline data accumulation will increase disk pressure.
|
||||
"""
|
||||
Loading…
Add table
Add a link
Reference in a new issue