YES, build this.
Cost-viable at $50-150/month. Adds a deep geopolitical analysis layer to the existing research ecosystem. NanoClaw provides autonomous orchestration, Supabase integration is trivial. Phase 1 (triage + alerts) deployable in 4-6 weeks, full pipeline over 12 months.

MVP Roadmap

PhaseTimelineMonthly CostCapabilityModels
Phase 1: Triage OnlyMonths 1-2$25/moRSS scan + alertsGemini Flash
Phase 2: Deep DiveMonths 3-4$75/moFull analysis on triggerFlash + Opus
Phase 3: Auto-RecMonths 5-6$125/moPosition sizing + intentFlash + GPT-5
Phase 4: AutonomousMonths 7-12$150/moFull pipelineFlash + GPT-5/Opus

Phase 1: Triage Only (Months 1-2, $25/mo)

Builds on top of the existing RSS pipeline. NanoClaw runs a triage agent that scans geopolitical feeds and generates WhatsApp alerts. No auto-trade. The goal is to validate that the triage model catches real events with acceptable precision. Monitor false positive rate and adjust scoring threshold.

Phase 2: Deep Dive on Trigger (Months 3-4, $75/mo)

High-score events trigger NanoClaw to spawn full deep-dive research agents in container isolation. Output: structured analysis JSON pushed to Supabase. Human reviews analysis before any trade decision. Validates research quality and signal accuracy.

Phase 3: Auto-Recommendation (Months 5-6, $125/mo)

Pipeline generates trade recommendations with position sizing. Intent is generated but requires human approval before execution. Tests the full pipeline end-to-end with a human gate.

Phase 4: Full Autonomous (Months 7-12, $150/mo)

NanoClaw runs the full pipeline autonomously. Findings flow to Supabase and are picked up by the weekly research combiner and portfolio rebalancer automatically. High-conviction findings (>85%) feed directly into rebalancing. Full autonomous mode only after Phase 3 proves reliable.

What Could Go Wrong

RiskImpactMitigation
False positivesTriage triggers on noise, wastes API spend on unnecessary deep divesPhase 1 validates triage precision before enabling deep dives. Adjust score threshold based on empirical false positive rate.
Analysis qualityLLM hallucination in research leads to bad trade recommendationReview agents (Wave 2 in Iran architecture) catch factual errors. Human gate in Phase 2-3 catches remaining issues. Objectivity audit agent verifies source diversity.
Latency50-80 min from event to trade may miss fast-moving opportunitiesAcceptable for geopolitical/macro events which play out over days to weeks. For genuinely time-critical events, the alert path (~24 min) enables manual fast action.
Cost creepMultiple events in one day exceed daily budgetHard daily cost cap in the orchestrator. Queue excess events for next-day processing. Priority scoring ensures the highest-impact event gets analyzed first.
Model degradationProvider model updates change output quality or formatStructured output schemas with validation. LLM Judge benchmarks detect quality drift. Model fallback chain (Opus -> GPT-5 -> Sonnet) if primary model degrades.

Next Steps