"Alerted.org

Job Title, Industry, Employer
City & State or Zip Code
20 mi
  • 0 mi
  • 5 mi
  • 10 mi
  • 20 mi
  • 50 mi
  • 100 mi
Advanced Search

Advanced Search

Cancel
Remove
+ Add search criteria
City & State or Zip Code
20 mi
  • 0 mi
  • 5 mi
  • 10 mi
  • 20 mi
  • 50 mi
  • 100 mi
Related to

  • AI Research Scientist - Safety Alignment Team

    Meta (Menlo Park, CA)



    Apply Now

    Summary:

    Meta is seeking AI Research Scientists to join the Safety Alignment team within Meta Superintelligence Labs, dedicated to advancing the safe development and deployment of superintelligent AI. Our mission is to pioneer robust safety alignment techniques that empower Meta’s most ambitious AI capabilities, ensuring billions of users experience our products and services securely and responsibly.

    Required Skills:

    AI Research Scientist - Safety Alignment Team Responsibilities:

    1. Design, implement, and evaluate novel safety alignment techniques for large language models and multimodal AI systems

    2. Create, curate, and analyze high-quality datasets for safety alignment

    3. Fine-tune and evaluate LLMs to adhere to Meta’s safety policies and evolving global standards

    4. Build scalable infrastructure and tools for safety evaluation, monitoring, and rapid mitigation of emerging risks

    5. Work closely with researchers, engineers, and cross-functional partners to integrate safety alignment into Meta’s products and services

    6. Lead complex technical projects end-to-end

    Minimum Qualifications:

    Minimum Qualifications:

    7. Bachelor's degree in Computer Science, Computer Engineering, relevant technical field, or equivalent practical experience

    8. PhD in Computer Science, Machine Learning, or a relevant technical field

    9. 3+ years of industry research experience in LLM/NLP, computer vision, or related AI/ML model training

    10. Experience as a technical lead on a team and/or leading complex technical projects from end-to-end

    11. Publications at peer-reviewed conferences (e.g. ICLR, NeurIPS, ICML, KDD, CVPR, ICCV, ACL)

    12. Programming experience in Python and hands-on experience with frameworks such as PyTorch

    Preferred Qualifications:

    Preferred Qualifications:

    13. Hands-on experience applying RL techniques (e.g., RLHF, PPO, DPO, GRPO, RLVF, reward modeling) to fine-tune large language models for safety and policy adherence

    14. Experience developing, fine-tuning, or evaluating LLMs across multiple languages and modalities (text, image, voice, video)

    15. Demonstrated experience to innovate in safety alignment, including custom guideline enforcement, dynamic policy adaptation, and rapid hotfixing of model vulnerabilities

    16. Experience designing, curating, and evaluating safety datasets, including adversarial and borderline prompt pairs for risk mitigation

    17. Experience with distributed training of LLMs (hundreds/thousands of GPUs), scalable safety mitigations, and automation of safety tooling

    Public Compensation:

    $177,000/year to $251,000/year + bonus + equity + benefits

    **Industry:** Internet

    Equal Opportunity:

    Meta is proud to be an Equal Employment Opportunity and Affirmative Action employer. We do not discriminate based upon race, religion, color, national origin, sex (including pregnancy, childbirth, or related medical conditions), sexual orientation, gender, gender identity, gender expression, transgender status, sexual stereotypes, age, status as a protected veteran, status as an individual with a disability, or other applicable legally protected characteristics. We also consider qualified applicants with criminal histories, consistent with applicable federal, state and local law. Meta participates in the E-Verify program in certain locations, as required by law. Please note that Meta may leverage artificial intelligence and machine learning technologies in connection with applications for employment.

     

    Meta is committed to providing reasonable accommodations for candidates with disabilities in our recruiting process. If you need any assistance or accommodations due to a disability, please let us know at [email protected].

     


    Apply Now



Recent Searches

  • Associate Director Technology Solution (Indiana)
  • Program Outreach Engineer (Virginia)
  • IBM Associate Partner SAP (New York)
[X] Clear History

Recent Jobs

  • AI Research Scientist - Safety Alignment Team
    Meta (Menlo Park, CA)
[X] Clear History

Account Login

Cancel
 
Forgot your password?

Not a member? Sign up

Sign Up

Cancel
 

Already have an account? Log in
Forgot your password?

Forgot your password?

Cancel
 
Enter the email associated with your account.

Already have an account? Sign in
Not a member? Sign up

© 2025 Alerted.org