Six years on the front lines of platform safety. Now on the frontier of AI — stress-testing models, evaluating outputs, and building toward a career in AI Governance.
I started reviewing content at scale in 2019 — flagging hate speech, coordinating abuse patterns, protecting users on social media and gaming platforms. That was my entry point into safety work: sitting with the worst of what people produce online, and making judgment calls thousands of times a day.
Six years on, I'm doing the same thing — but for AI systems. I adversarially test large language models, probing the gaps between what a model says it won't do and what it actually does when you push the right way. The instincts built from years of human content review turn out to be exactly what red teaming needs.
What makes my path unusual is the academic lens I bring. Two degrees in psychology and history mean I don't just ask how a system fails — I ask why. I think about intent, context, and the human behaviour underneath the content. That combination of operational experience and analytical thinking is what I'm building a career in AI Governance on.
From content moderation to LLM training - every role added a new layer to how I think about safety, policy, and human behaviour online.
A toolkit built across six years of safety operations — from human content review to adversarial AI testing.
Two degrees that shaped how I think about human behaviour — and three certifications that sharpened the technical edge.
Safety work is judgment work. These are the principles I bring to every case — whether it's a piece of harmful content or an LLM guardrail under pressure.
The things that inform how I think — not just what I do.
Open to remote roles in Trust & Safety, LLM Training, and AI Governance. If you're building safer AI systems, I'd like to help.