Insights
Writing about what I’m building and what breaks along the way. All posts live on my blog.
Featured
1,500 Lines, One Commit
I built a full custom charting system, then deleted 1,923 lines when TradingView's widget turned out to be simpler and better.
Agent Memory at Scale
Vector databases alone broke immediately. What worked: Postgres + Qdrant + Redis with three explicit memory types.
The Model That Never Made a Trade
I wrote a PPO reinforcement learning trading agent, then shelved it when LLMs with tool-calling shipped the same week.
All Posts
1,500 Lines, One Commit Tools
Feb 25, 2026
I built a full custom charting system for my AI financial analysis platform – drawing tools, per-asset persistence, keyboard shortcuts, context providers, Bollinger Band colors. Weeks of work. Then I looked at TradingView’s embeddable widget and deleted 1,923 lines in one commit. The deletion turned out to be the most productive commit on the project: it forced me to rethink how the AI sees chart data, and the answer (capture what the user actually sees, send it to a vision model) was simpler and better than what I’d been building toward.
Building Real-Time AI Financial Analysis: Why We Use TypeScript AND Python Architecture
Feb 8, 2026
We rebuilt our financial analysis platform twice before getting the architecture right. First attempt was Python for everything – worked until frontend developers couldn’t touch the backend and we were catching type errors in production. Second attempt was TypeScript for everything – felt great until the AI/ML ecosystem gaps started costing us weeks. The version that finally scaled splits TypeScript (API layer, real-time WebSockets, type safety at user boundaries) from Python (LLM orchestration, pandas, TA-Lib, the entire ML toolkit), connected via gRPC. PRs per day went up 4x and compute costs dropped 30%.
The Model That Never Made a Trade AI
Feb 22, 2026
I wrote a PPO reinforcement learning trading agent – defined the gymnasium environment, set up the observation space, imported PyTorch and stable-baselines3. Then I shelved it. Training an RL agent on financial data is a real research project: simulated trading infrastructure, careful reward shaping so the model doesn’t learn to just never trade, and compute to iterate. Meanwhile, LLMs with tool-calling shipped the same week – no training data, 2GB lighter Docker image, useful analysis out of the box.
The EU Doesn’t Kill Innovation. The US Regulatory Mess Might. Opinion
Feb 18, 2026
The “Europe is a graveyard for tech” story is mostly wrong. I’ve shipped products across AI, fintech, and crypto on both sides of the Atlantic. The EU’s MiCA gives crypto builders one license framework across 27 countries. The US has 50 separate state licensing regimes and a habit of regulating by lawsuit – you find out the rules after someone sues you.
Software Is Not Dead – It’s Evolving Opinion
Feb 16, 2026
Mark Cuban said “software is dead” because AI will customize everything. He’s wrong about the conclusion, even if the direction is right. AI systems run inside containers, API gateways, databases, logging pipelines – all traditional software engineering. The real shift isn’t the death of software; it’s the move from static SaaS to adaptive systems.
ObsiTUI: Why I Built a Terminal UI for Obsidian Tools
Feb 15, 2026
I spend my day in the terminal and my knowledge base lived in Obsidian’s Electron app. Every note lookup meant alt-tabbing, grabbing the mouse, switching mental contexts – 30-60 seconds each time, dozens of times a day. So I built ObsiTUI: a terminal client for Obsidian with real vim modal editing, AI-powered RAG search, and a pure functional editor where every operation returns a new immutable state. It has 301 tests.
What We Learned Building Agent Memory at Scale AI
Feb 15, 2026
Our AI agents couldn’t remember anything useful across sessions. We tried vector databases alone – broke immediately because semantic similarity doesn’t understand importance or recency. What worked: a hybrid of Postgres, Qdrant, and Redis, with three explicit memory types – user-specific, session-specific, and shared pools that let agents learn from each other.
AI-Powered DevOps Orchestration Architecture
Jan 22, 2026
Traditional autoscaling is crude – CPU hits 70%, spin up more servers, half of them idle while you pay for them. We’re replacing threshold-based scaling with ML models that predict traffic spikes before they happen. Prophet and TensorFlow on Prometheus metrics, triggered from Bitbucket Pipelines.
Boosting Our Financial AI Project with LangChain AI
Oct 3, 2025
LangChain let us stop rewriting code every time we wanted to test a different LLM. We swap between GPT-4, Claude, Llama, and Grok with minimal code changes, testing each on specific financial tasks – market summaries, chatbot responses, portfolio optimization.
Unleashing AI Trading Potential with MCP AI
Nov 2025
LLMs are good at reasoning about markets but bad at accessing live data. Anthropic’s Model Context Protocol fixes this with a standard way to connect models to external tools – one server per data source, one client per AI app, communicating over JSON-RPC.
From Code to Assets: Revolutionizing Trading with AI AI
Sep 2025
Where we started. At Hatchworks VC, our team moved from experimental ML models to agentic AI systems for trading. We fine-tune LLMs on unstructured data for real-time market sentiment, and we’re building custom RL models for portfolio optimization in volatile markets.