Opening. This role exists to develop and implement novel finetuning techniques using synthetic data generation and advanced training pipelines to train models to have better alignment properties including honesty, character, and harmlessness.
What you'll do
As a Research Scientist/Engineer on the Alignment Finetuning team at Anthropic, you'll lead the development and implementation of techniques aimed at training language models that are more aligned with human values: that demonstrate better moral reasoning, improved honesty, and good character.
- Develop and implement novel finetuning techniques using synthetic data generation and advanced training pipelines
- Use these to train models to have better alignment properties including honesty, character, and harmlessness
What you need
- Develop processes to help automate and scale the work of the team
- Collaborate across teams to integrate alignment improvements into production models
Why this matters
One paragraph about career impact and value.