The GenAI Evaluations Foundations team is looking to hire a Research Scientist with Vision experience for the Safety Evaluations Dev team. Safety Evaluations are key to help us both internally align as well as externally share on how the LLM is being safe in its responses to adversarial or unsafe prompts. When our models are safe, we can feel confident and comfortable open-sourcing our models for other developers and our internal product teams to use.
Evaluations & benchmarks are what steer AI progress, as we can inject them in all stages of the model training. The sooner we can catch issues, the faster we can fix things resulting in us saving millions of dollars and compute cycles while avoiding harming Meta’s reputation in the world.
As Llama further builds on its image understanding and image generation modalities, we are looking to hire a Research Scientist who is passionate about Safety and has experience designing evaluations and dataset for Computer vision models.
Research Scientist, GenAI Safety Evaluations - Computer Vision Responsibilities
- Design and implement datasets to evaluate our LLMs on safety, with a focus on Vision.
- Adapt standard machine learning methods to best exploit modern parallel environments (e.g. distributed clusters, multicore SMP, and GPU).
- Work with a large and globally distributed team across multiple functions to understand the needs and align on goals and outcomes.
- Play a significant role in healthy cross-functional collaboration.
Minimum Qualifications
- Bachelor's degree in Computer Science, Computer Engineering, relevant technical field, or equivalent practical experience.
- PhD in Computer Science, Computer Engineering, relevant technical field
- 3+ year(s) of work experience in a university, industry, or government lab with emphasis on AI Research in machine learning, deep learning, and computer vision.
- Programming experience in Python and experience with frameworks such as PyTorch.
- Exposure to architectural patterns of large scale software applications.
- Domain-Relevant Research Publications accepted at peer-reviewed AI conferences (e.g., NeurIPS, CVPR, ICML, ICLR, ICCV, and ACL)
Preferred Qualifications
- Experience working with Safety or related areas.
- Direct experience in building evals for generative AI and LLM research
- First author publications at peer-reviewed AI conferences (e.g., NeurIPS, CVPR, ICML, ICLR, ICCV, and ACL).
For those who live in or expect to work from California if hired for this position, please click
here for additional information.
About Meta
Meta builds technologies that help people connect, find communities, and grow businesses. When Facebook launched in 2004, it changed the way people connect. Apps like Messenger, Instagram and WhatsApp further empowered billions around the world. Now, Meta is moving beyond 2D screens toward immersive experiences like augmented and virtual reality to help build the next evolution in social technology. People who choose to build their careers by building with us at Meta help shape a future that will take us beyond what digital connection makes possible today—beyond the constraints of screens, the limits of distance, and even the rules of physics. Meta is committed to providing reasonable support (called accommodations) in our recruiting processes for candidates with disabilities, long term conditions, mental health conditions or sincerely held religious beliefs, or who are neurodivergent or require pregnancy-related support. If you need support, please reach out to
accommodations-ext@fb.com. $१,७७,०००/year to $२,५१,०००/year + bonus + equity + benefits
Individual compensation is determined by skills, qualifications, experience, and location. Compensation details listed in this posting reflect the base hourly rate, monthly rate, or annual salary only, and do not include bonus, equity or sales incentives, if applicable. In addition to base compensation, Meta offers benefits. Learn more about
benefits at Meta.