Logo
Karkidi

Policy Design Manager, Child Safety and Emotional and Psychological Harm

Karkidi, San Francisco, California, United States, 94199


As a Trust and Safety Policy Manager, you will be responsible for helping develop and manage policies for our products and services. Safety is core to our mission and as an early member of the team, you’ll help shape policy creation and development so that our users can safely interact with and build on top of our products in a harmless, helpful and honest way.About Anthropic

Anthropic’s mission is to create reliable, interpretable, and steerable AI systems. We want AI to be safe and beneficial for our users and for society as a whole. Our team is a quickly growing group of committed researchers, engineers, policy experts, and business leaders working together to build beneficial AI systems.Responsibilities:

Draft new policies that help govern the responsible use of our models for emerging capabilities and use cases.Conduct regular reviews of existing policies to identify and address gaps and ambiguities.Iterate on and help build out our comprehensive harm framework.Develop deep subject matter expertise in a particular policy area.Update our policies based on feedback from our enforcement team and edge cases that you will review.Educate and align internal stakeholders around our policies and our overall approach to product policy.Partner with internal and external researchers to better understand our product’s limitations and risks and adapt our policies based on such.Work closely with enforcement and detection teams to identify policy gaps based on violations and edge cases.Keep up to date with new and existing AI policy norms and standards, and use these to inform our decision-making on policy areas.You may be a good fit if you:

Have experience drafting or updating product and/or user policies.Have experience aligning product policy decisions between diverse sets of stakeholders, such as Product, Engineering, Security, and Legal teams.Understand the challenges that exist in developing and implementing product policies at scale, including in the content moderation space.Love to think creatively about how to use technology in a way that is safe and beneficial, and ultimately furthers the goal of advancing safe AI systems.Have a passion for or interest in artificial intelligence and ensuring it is developed and deployed safely.Annual Salary (USD)

The expected salary range for this position is $200k-$240k.Logistics

Location-based hybrid policy: Currently, we expect all staff to be in our office at least 25% of the time.Deadline to apply: None. Applications will be reviewed on a rolling basis.US visa sponsorship: We do sponsor visas! However, we aren't able to successfully sponsor visas for every role and every candidate; operations roles are especially difficult to support. But if we make you an offer, we will make every effort to get you into the United States, and we retain an immigration lawyer to help with this.Compensation and Benefits

Anthropic’s compensation package consists of three elements: salary, equity, and benefits. We are committed to pay fairness and aim for these three elements collectively to be highly competitive with market rates.Equity - On top of this position's salary (listed above), equity will be a major component of the total compensation. We aim to offer higher-than-average equity compensation for a company of our size, and communicate equity amounts at the time of offer issuance.US Benefits

Optional equity donation matching at a 3:1 ratio, up to 50% of your equity grant.Comprehensive health, dental, and vision insurance for you and all your dependents.401(k) plan with 4% matching.22 weeks of paid parental leave.Unlimited PTO – most staff take between 4-6 weeks each year, sometimes more!Stipends for education, home office improvements, commuting, and wellness.Fertility benefits via Carrot.Daily lunches and snacks in our office.Relocation support for those moving to the Bay Area.

#J-18808-Ljbffr