I needed a place to share my thoughts. What I'm thinking about now, what I thought about in the past, and what we could think about together.
I study AI safety and security at Microsoft on the AI Red Team. I'm focused on what happens as people welcome AI into their most personal contexts, and on how that closeness can be better researched and understood so we can get it right.
Also, you should meet my dog, Sparky. He walks around looking like he's about to have a thought but never quite finishes it. This might be a good place for those thoughts too.
Eugenia Kim · AI Safety & Security Researcher
Sparky · Golden Retriever · Thinks about things sometimes
"Nobody asked for anything. I just did sit, spin, down, up, jump all at once and got a high-reward-value treat."
"If you have a treat, I can do every trick in the book. No treat? Never heard of the command 'sit' in my life."
The apartment buzzer. Zero-knowledge threat. Signal arrives without warning, actor is never visible. Sometimes it's a cardboard box. Sometimes Eugenia ignores it and I never find out. Threat level unassessable. I bark every time. This is the only responsible policy.
The vacuum cleaner. Unpredictable schedule. Sometimes morning, sometimes after lunch. If I carefully destroy a plush toy and disperse its fluff, the vacuum removes all of it. High-frequency, indiscriminate, broadly dispersed harm. Affects me and toy. No known mitigation.
The new puppy in the building. Very cute. Fun to play with. But I used to be the cute one. Monitoring a gradual redistribution of attention and treat resources. Not confirming threat. Not ruling it out. Logging for now.
Leading all foundational research on frontier-model behavior related to psychosocial harms. Designing and engineering open-source automated red-teaming infrastructure. Previously SDE2 building tools for red teaming operations including PyRIT.
Algorithmic analysis of news and social media coverage of mental health topics. Co-authored research with CDC researchers on suicide framing in media.
Analyzed age bias in facial emotion recognition systems. First-authored publication on AI bias mitigation at AIES '21.
Undergraduate research in organic electronics and self-assembly methodologies.
Interested in collaborating on AI safety, red teaming, or psychosocial harms research? Always open to interesting problems.
Email meFind me on these platforms.