About the Role
To help API users monitor and prevent unwanted use cases, we developed the moderation endpoint, a tool for checking whether content complies with OpenAI's content policy. Developers can thus identify content that our content policy prohibits and take actions (e.g. block it). We seek a Machine Learning Engineer to help design and build a robust pipeline for data management, model training and deployment to enable a consistent improvement on the Moderation model.
In this role, you will:
Design, develop and maintain a robust and scalable data management pipeline and set up standards for versioning and data quality control. The pipeline should be able to handle data relabeling requests due to content policy changes.
Build a pipeline for automated model training, evaluation and deployment, including active learning process, routines for calibration and validation data refresh etc.
Work closely with stakeholders from product, engineering, content policy on a long-term improvement over the moderation models, for both external release and internal use cases across a variety of projects on model safety.
Research on the latest techniques and methods in deep learning and natural language processing to improve the moderation model across a collection of unwanted content categories.
Experiment on data augmentation and data generation methods to enhance the diversity and quality of training data.
Experiment and design an effective red-teaming pipeline to examine the robustness of the model and identify areas for future improvement.
Conduct open-ended research to improve the quality of collected data, including but not limited to, semi-supervised learning and human-in-the-loop machine learning.
You might thrive in this role if you:
Have 3+ years industry experience as a Machine Learning Engineer or Software Engineer, working on building data pipelines, training and deploying machine learning models in production on a daily basis.
Care deeply about AI safety and passionate about building the best deep learning empowered moderation model to effectively detect unwanted content.
Have a strong belief in the criticality of high-quality data and are highly motivated to work with the associated challenges.
Have experience working in large distributed systems, deep learning or/and natural language processing is a big plus.
Love working with a team.
OpenAI is an AI research and deployment company dedicated to ensuring that general-purpose artificial intelligence benefits all of humanity. We push the boundaries of the capabilities of AI systems and seek to safely deploy them to the world through our products. AI is an extremely powerful tool that must be created with safety and human needs at its core, and to achieve our mission, we must encompass and value the many different perspectives, voices, and experiences that form the full spectrum of humanity.
At OpenAI, we believe artificial intelligence has the potential to help people solve immense global challenges, and we want the upside of AI to be widely shared. Join us in shaping the future of technology.
Compensation, Benefits and Perks
The annual salary range for this role is $200,000 – $370,000. Total compensation also includes generous equity and benefits.
Medical, dental, and vision insurance for you and your family
Mental health and wellness support
401(k) plan with 4% matching
Unlimited time off and 18+ company holidays per year
Paid parental leave (20 weeks) and family-planning support
Annual learning & development stipend ($1,500 per year)
We are an equal opportunity employer and do not discriminate on the basis of race, religion, national origin, gender, sexual orientation, age, veteran status, disability or any other legally protected status. Pursuant to the San Francisco Fair Chance Ordinance, we will consider qualified applicants with arrest and conviction records.