Skip to main content

Leverage Record: March 5, 2026

AI Time Record

About the author: I'm Charles Sieg, a cloud architect and platform engineer who builds apps, services, and infrastructure for Fortune 1000 clients through Vantalect. If your organization is rethinking its software strategy in the age of AI-assisted engineering, let's talk.

About These Records
These time records capture personal project work done with Claude Code (Anthropic) only. They do not include work done with ChatGPT (OpenAI), Gemini (Google), Grok (xAI), or other models, all of which I use extensively. Client work is also excluded, despite being primarily Claude Code. The actual total AI-assisted output for any given day is substantially higher than what appears here.

Yesterday was one of the highest-volume days I have recorded. Thirty distinct tasks across the full spectrum of the work: domain specification generation, full-stack application development, patent portfolio maintenance, cloud platform engineering, mobile app development, and business planning. The numbers tell the story.

Task Log

# Task Human Est. Claude Supv. LF SLF
1 AI assistant tool chain extension: new configuration endpoints, system prompt updates, sidebar and cache fixes 16h 32m 5m 30x 192x
2 Resume upload modal and calibration enhancements across 15 files (engine, demo client, desktop client) 16h 7m 5m 137x 192x
3 Open-source diagramming library: fix sibling alignment double-shift bug, extend dogleg clearance, add 4 tests 8h 19m 5m 25x 96x
4 Full patent portfolio audit (11 applications): fix 16 claim back-references, 8 unreferenced figures, update documentation 24h 35m 5m 41x 288x
5 Update filing dates across PDF generation scripts and regenerate all 44 PDFs (11 continuation + 33 branded) 4h 25m 5m 10x 48x
6 Fix PDF structural warnings: add Ghostscript re-distill step to generation pipeline, regenerate 11 PDFs 3h 12m 5m 15x 36x
7 Cloud platform: costs, compliance, org-scan, resource-types, navigation; 14 new files, 9 modified 40h 25m 5m 96x 480x
8 Create admin dashboard repository (21 files, React/Vite) 8h 8m 5m 60x 96x
9 Cloud platform: reports, automations, seed data, and test system 16h 12m 5m 80x 192x
10 Mandatory multi-method MFA implementation (TOTP + email + SMS) for authentication service 16h 5m 5m 192x 192x
11 Domain specification validation fixes: expand 4 specs from 55-58 to 62 leaves, fix verb issues 2h 8m 5m 15x 24x
12 Build full issue-tracking application (121 files: FastAPI backend + React kanban board) 40h 21m 5m 114x 480x
13 Product documentation: README, requirements, and design documents 12h 8m 5m 90x 144x
14 Create 3 HR/ERP certification domain specification files 8h 18m 5m 27x 96x
15 UI standardization: replace toast library, standardize charting, create style guide, retrofit across 3 applications 16h 25m 5m 38x 192x
16 Admin console: 8 service views + MCP server with 96 tools 120h 45m 5m 160x 1440x
17 Fix 9 domain specs (expand to 60 leaves) + fix verb/word issues in 11 pre-existing specs 4h 8m 5m 30x 48x
18 Bug report button + issue tracker enhancements + MCP server integration 24h 35m 5m 41x 288x
19 Domain specification continuation: create 2 missing specs, fix 2 uniform trees 4h 12m 5m 20x 48x
20 Domain specification cleanup: UUID canonicalization, duplicate removal for 58 total specs 4h 8m 5m 30x 48x
21 Batch domain specification creation: ~487 specs across 13+ certification vendors with validation 1500h 120m 5m 750x 18000x
22 Vector database integration for RAG pipeline (Milvus implementation + verification) 4h 15m 5m 16x 48x
23 Mobile app gap analysis implementation: 11 gaps across 7 phases (calibration, session config, adaptation, knowledge map, lessons, exam API, progress) 120h 45m 5m 160x 1440x
24 Fix drafting issues across 11 patent applications (language precision, undefined variables, over-narrowing) 16h 25m 5m 38x 192x
25 AI assistant: clear chat button + ASCII diagram tool (6 files across 2 repos) + global config + client fixes 6h 50m 5m 7x 72x
26 Implement chat panel in desktop client for feature parity with web demo 8h 12m 5m 40x 96x
27 Conversational assistant subsystem: full-stack implementation across engine + 2 clients + reference architecture 120h 35m 5m 206x 1440x
28 Voice interaction and hands-free mode for mobile app (3 new files + 10 modified) 16h 25m 5m 38x 192x
29 Draft business plan, marketing plan, valuation/funding model, and pitch deck for EdTech startup 80h 35m 5m 137x 960x
30 Comprehensive exam market analysis document for EdTech planning 8h 8m 5m 60x 96x

Legend: Human Est. = estimated human-equivalent time. Claude = wall-clock minutes for Claude to complete. Supv. = minutes I spent writing the prompt. LF = leverage factor (human time / Claude time). SLF = supervisory leverage factor (human time / my time).

Aggregate Statistics

Metric Value
Total tasks 30
Total human-equivalent hours 2,263
Total Claude minutes 738 (12.3 hours)
Total supervisory minutes 150 (2.5 hours)
Total tokens consumed ~4,680,000
Weighted average leverage factor 184.0x
Weighted average supervisory leverage factor 905.2x

Analysis

The dominant task was domain specification generation. Record 21 alone accounts for 1,500 human-equivalent hours: creating 487 structured certification domain specifications across 13+ vendors (IBM, Oracle, Salesforce, VMware, EC-Council, and others). Each specification required researching the certification exam blueprint, structuring knowledge domains into hierarchical trees with 60-80 leaf nodes, generating seed question chains, and validating the output against schema constraints. A single human domain expert would need weeks per vendor. Claude produced all 487 in two hours of wall-clock time, yielding a 750x leverage factor. The supervisory leverage on that task (18,000x) reflects the fact that one five-minute prompt generated the entire batch.

The second tier of high-leverage work was full-stack application development. Building the conversational assistant subsystem (206x), the admin console with 96 MCP tools (160x), and the mobile app gap analysis implementation (160x) each represented multiple weeks of human-equivalent engineering compressed into under an hour of Claude execution.

The lowest leverage factors appeared on tasks involving iterative tool use and external process dependencies: the PDF regeneration pipeline (10x) required waiting on Ghostscript and qpdf processes, and the multi-repo configuration work (7x) involved chasing down path inconsistencies across several codebases. These tasks have a higher ratio of "waiting on the machine" to "thinking," which compresses the leverage factor.

The weighted average supervisory leverage of 905x means that for every minute I spent writing prompts, I received 905 minutes of equivalent human engineering output. Put differently, my 2.5 hours of supervisory time yesterday produced what would have taken a team of engineers roughly 56 work weeks.

Let's Build Something!

I help teams ship cloud infrastructure that actually works at scale. Whether you're modernizing a legacy platform, designing a multi-region architecture from scratch, or figuring out how AI fits into your engineering workflow, I've seen your problem before. Let me help.

Currently taking on select consulting engagements through Vantalect.