Your AI Training Cluster Thirsty? Let's Talk Water.
We ran the numbers: A 10k H100 cluster can consume 2 million gallons of water a month. Here is the math and the engineering fix.
"Add this article to our feed."
npm run url-news "https://techcrunch.com/article-url"That's it. Article extracted, analyzed, tagged, and added.
We initially tried two extremes:
The solution was a hybrid architecture.
Before we waste a single token, we process the content locally.
We built a robust extractor that runs right on the machine. It handles:
og:title, authors, dates, and site names using standard meta tags.Once we have clean, high-signal text, then we bring in the heavy guns. We pass the cleaned JSON to an LLM (Claude or OpenAI) for the tasks that actually require intelligence:
1. Cost Control
By cleaning the HTML locally, we reduce the token count by 60-80% before the API call. We pay to process information, not <div> tags.
2. Speed
Local relevance scoring means we can discard low-value URLs in milliseconds without network latency.
3. Reliability
If the LLM is down or hallucinates, we still have the locally extracted title, date, and raw content. The system degrades gracefully.
We use a "waterfall" strategy. Try the most reliable method (Open Graph tags); if missing, fall back to semantic HTML (<article>); if missing, use heuristics (largest text block). This ensures we get something usable from almost any site.
We maintain a weighted dictionary of domain-specific terms ("PPA", "interconnection", "H100"). An article must cross a point threshold to be considered "intelligence." This simple filter saves us from filling our database with generic tech news.
You don't need AI for everything.
The most effective AI systems are often 20% AI and 80% solid engineering. By letting code do what code does best (scraping, filtering, formatting), we free up the AI to do what it does best (reasoning and synthesis).
See our Intelligence Feed in action at /intelligence-feed. The source code for our extractor is available in our repository.
We ran the numbers: A 10k H100 cluster can consume 2 million gallons of water a month. Here is the math and the engineering fix.
Traditional SaaS is too slow for energy markets. We pivoted to 'Autonomous Organization as a Service'—software that works while you sleep.
Giving an agent 30 tools costs $0.45 per run. We implemented a 'Code-First Skills' pattern to drop that to $0.003.
Grid interconnection is the #1 bottleneck for AI. Google X's Tapestry project is trying to virtualize the grid to fix it.
News tells you what happened yesterday. Markets tell you what will happen tomorrow. We built an agent to trade on the difference.
Starting August 2025, mandatory environmental reporting kicks in for AI models. Most CTOs are completely unprepared.
We forced our AI agents to fight. The 'Bull' vs. The 'Bear'. The result was better decisions than any single model could produce.
Installed capacity is a vanity metric. LCOE is the only number that levels the playing field between solar, gas, and nuclear.
Grid carbon intensity varies by 3x throughout the day. We built a scheduler that pauses AI training when the grid is dirty.