Research Scientist, Alignment & Safety

Duration: Permanent

Payrate: 180k-200k

Technical/Functional Skills:

  • Strong understanding of machine learning principles, especially in the context of LLMs.
  • Knowledgeable about different LLM approaches for aligning LLMs (SFT, RLHF, etc.).
  • Can own and pursue a research agenda, including choosing impactful research problems and autonomously carrying out projects.
  • Fluent in at least one statistical programming language such as Python (preferred) or R.
  • Demonstrated background in collecting data from human participants (e.g., surveys, experiments) with knowledge about data quality, data validity, etc.
  • Strong verbal and written communications skills with the ability to work effectively across internal and external organizations and virtual teams.
  • PhD or advanced degree in computer science, machine learning, cognitive science, psychology, economics, or similar (preferred).

Skills:

  • ML
  • LLMs
  • Python
  • R

Experience Required:

  • Strong understanding of ML principles, especially in the context of LLMs. Fluent in at least one statistical programming language such as Python (preferred) or R.

Roles & Responsibilities:

  • Research how to best use techniques like Supervised Fine Tuning (SFT) and Reinforcement Learning with Human Feedback (RLHF) to align LLMs.
  • Work with technical and non-technical stakeholders to align LLMs for specific use cases.
  • Research approaches to ensure LLMs are safe, preventing the generation of harmful content and maintaining alignment with societal norms.
  • Research scalable oversight mechanisms that enable efficient monitoring and control of LLMs as they grow in size and complexity, ensuring consistent alignment with predefined objectives.