[portfolio] Portfolio Dashboard - 2026-03-09 #20179
Closed
Replies: 1 comment
-
|
This discussion was automatically closed because it expired on 2026-03-10T09:26:20.499Z.
|
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
This week's analysis covers 287 workflow runs across 78 unique workflows over a ~2-day window (2026-03-07 to 2026-03-09), with $45.62 total cost measured from 24 runs that have full token/cost instrumentation. The headline finding: 48% of all runs failed — a significant operational health issue concentrated in a handful of always-failing workflows.
Visual Summary
Cost Trends
Key Insights:
Top Cost Drivers
Top 3 workflows account for ~42% of tracked cost:
go-logger-enhancement.md— $8.16/period (2 runs, 0% failure) ✅copilot-session-insights.md— $5.29/period (2 runs, 0% failure) ✅sergo-serena-go-expert.md— $4.01/period (2 runs, 0% failure) ✅Failure Analysis
48% of all runs failed — driven by a small set of 100%-failing workflows:
Overall Health
💰 Cost Reduction Opportunities
Priority: Fix/Disable 100%-Failing Workflows to recover CI minutes and improve signal quality
Since most failing workflows have no cost instrumentation, direct dollar savings are limited, but recovering CI capacity and reducing noise is high value.
Strategy 1: Disable 100%-Failing Workflows (Critical)
These workflows have never succeeded in the observed period and should be investigated and either fixed or disabled:
Action: Audit each workflow. For event-driven workflows (Issue Monster, AI Moderator), check trigger configuration. For smoke tests (Smoke Codex), check if the engine/environment is broken.
Strategy 2: Investigate High-Cost Workflows for Frequency Tuning
These workflows have full cost data and run on schedule. All currently healthy (0% failure), but worth reviewing frequency vs. value:
Highest priority:
Daily Code Metrics and Trend Tracking Agent— $2.04 cost with 100% failure rate. Fix or disable immediately.Frequency review: Workflows running daily that cost >$2/run could save 50-75% by switching to weekly. Example: if
Go Logger Enhancementruns daily at ~$4/run → switching to weekly saves ~$84/month.Strategy 3: Investigate High-Volume Event Workflows
Issue Monster(65 runs) andAI Moderator(22 runs) are event-driven and both failing 100%. These are high-volume and completely broken.Issue Monster — 65 failures is the single largest contributor to the 48% failure rate. Fixing this one workflow alone would drop the portfolio failure rate from 48% to ~26%.
AI Moderator — 22 failures, likely similar root cause.
Recommended action: Check error logs for these two workflows first.
🎯 Priority Actions
CRITICAL — Fix or disable
Issue Monster: 65 failed runs is the largest single source of failures; fixing it alone cuts overall failure rate nearly in half. Check.github/workflows/issue-monster.mdfor trigger and permission issues.CRITICAL — Fix or disable
AI Moderator: 22 failed runs, likely same root cause as Issue Monster (event-driven workflows with misconfigured triggers or auth).HIGH — Disable
Daily Code Metrics and Trend Tracking Agent: $2.04/run with 100% failure rate = pure waste. Either fix the underlying issue or disable until repaired.HIGH — Audit
Smoke Codex(14 failures, 100%): If Codex engine is broken/unavailable, all Codex smoke tests will continue failing indefinitely.MEDIUM — Review frequency of
Go Logger Enhancement($8.16/2 days): If this runs daily at that cost, monthly spend is ~$122 for a single workflow. Verify business need for daily cadence.📈 Data Quality
Methodology: Analysis based on pre-downloaded workflow execution data from
gh aw logs. Costs calculated from actual token usage for instrumented runs. Failure rates computed from all 287 runs regardless of cost instrumentation.References:
Beta Was this translation helpful? Give feedback.
All reactions