I'm working with a cutting-edge AI safety start-up that's hiring a
Research Engineer (AI Alignment & Safety)
.
What You'll Do
* Evaluate advanced AI systems and detect potential risks (e.g. deceptive behaviors)
* Work on interpretability research to uncover how models really work
* Build tools that turn research into scalable, production-ready evaluations
What we're looking for:
* Strong Python + ML/NN background
* Experience in AI safety, alignment, or interpretability research
* Ability to write clean, production-quality code
* Curiosity, analytical mindset, and strong communication
Great opportunity to work with leading researchers and have a real impact on safe AI development. Competitive pay + benefits included.