Senior ML Engineer - AI Safety & Evaluation
We’re building a future where AI systems are not only powerful but safe, aligned, and robust against misuse. Our team focuses on advancing practical safety techniques for large language models (LLMs) and multimodal systems—ensuring these models remain aligned with human intent and resist attempts to produce harmful, toxic, or policy-violating content.
We operate at the intersection of model development and real-world deployment, with a mission to build systems that can proactively detect and prevent jailbreaks, toxic behaviors, and other forms of misuse. Our work blends applied research, systems engineering, and evaluation design to ensure safety is built into our models at every layer.
Position Overview
As a Senior ML Engineer on the AI Safety & Evaluation team, you will contribute to the development, deployment, and monitoring of model-level safety components in production environments. This includes building APIs and infrastructure to integrate safety checks, running evaluations under adversarial scenarios, and deploying models and safety modules in scalable and production-ready environments.
You’ll work closely with ML engineers, infrastructure teams, and product safety leads to ensure that our models are not only performant—but robust, auditable, and secure in real-world use cases.
What You’ll Do
Nice to Have
Compensation: up to $155K USD
Read Full Description