I'm an end-to-end AI builder focused on evaluations, benchmarks, and verification β making AI systems more rigorous and trustworthy. Currently at Waypoint, working on evaluations, knowledge graphs, and AI strategy.
My background spans neuroscience research (Yale, SRI/Stanford), data science, and software engineering. I believe evals are the weak link in AI development, and I'm working to change that.
- π€ Speaking at meetups and conferences on AI evaluation practices ("The Eval Flywheel", "Evals, Benchmarks, and Guardrails")
- π¨πΏ Organizing the Evals.cz meetup in Prague
- ποΈ Co-hosting the Data Talk and AI ta Krajta podcasts
- π©βπ« Teaching Data & AI courses at Czechitas
- Don't Write Evals for Fast-Moving Systems (2026)
- Clobsidian in Detail: Cross-Source Personal Infrastructure (2026)
- Clobsidian, and other winter experiments with Claude Code (2026)
See my blog for a full list of articles.
- Evals, Benchmarks, and Guardrails: A Pythonista's Guide to Not Mixing Them Up β PyData Prague, Feb 2026
- When (& How) to Start Writing Evals β Evals.cz Meetup #1, Feb 2026
- Simultaneous Vibe Coding β TopMonks CaffΓ©, Jan 2026
- Pydantic, Everywhere, All at Once β EuroPython Prague, Jul 2025




