If an employer mentions a salary or salary range on their job, we display it as an "Employer Estimate". If a job has no salary data, Rise displays an estimate if available.
At Anthropic, we're on a mission to create AI systems that are reliable, interpretable, and steerable, and we're looking for a Safeguards Policy Analyst to join our vibrant team in San Francisco or New York City. In this exciting role, you’ll get to design and execute enforcement workflows for our innovative products and services with a strong emphasis on identifying and mitigating potential harmful uses. As a valued member of our user Integrity and Authenticity team, you’ll focus on enhancing existing policies and expanding workflows related to integrity and authenticity. You'll also have the chance to partner closely with Product, Engineering, and Data Science teams to optimize detection models—essentially shaping how our community interacts with AI. While the role may require you to navigate some explicit content, your work will directly contribute to making sure our offerings are safe, beneficial, and honest for users. Excitingly, you'll have the chance to engage with external experts to gather feedback that informs our policies, ensuring we’re at the forefront of AI safety best practices. If you have experience in policy enforcement, excellent communication skills, and a knack for working effectively in a dynamic environment, we can’t wait to meet you! Come help us build a future where AI is a safe and supportive part of our users' lives.
Anthropic is an AI startup public-benefit company dedicated to AI safety and research, aiming to develop dependable, interpretable, and controllable AI systems. The company was was founded by former members of OpenAI in 2021.
526 jobsSubscribe to Rise newsletter