Data engineering
Scalable lakehouse and warehouse patterns, Spark and distributed batch, plus operational rigor for production datasets.
Data & ML infrastructure
Data engineer building reliable data platforms, ML pipelines, and real-time analytics—where scale meets clarity.
Open to conversations on data platforms, ML in production, and pragmatic system design.
I design and build data infrastructure that stays understandable under load: clear contracts, observable pipelines, and systems that teams can evolve.
I care about the gap between a prototype notebook and something that runs every day in production—scheduling, failure modes, cost, and the human side of operating platforms.
If you’re working on similar problems, the notes below are where I document what actually worked (and what didn’t).
Hands-on work across the stack—from streaming and batch to model serving and observability.
Scalable lakehouse and warehouse patterns, Spark and distributed batch, plus operational rigor for production datasets.
Training-to-serving paths, batch and online inference, and APIs that stay maintainable as models and traffic evolve.
Operators, workloads on K8s, and automation so teams can ship data products without fragile one-off scripts.
Streaming ingestion, low-latency paths, and tooling so stakeholders see fresh signal—not yesterday’s snapshot.
Small browser experiments—good for a short break between deep work.
More gamesSelected projects and long-form notes from the blog.
If my writing or tools helped you, TRC20 donations are welcome.