About the role
As a data scientist in the Responsible Development and Innovation (ReDI) team, you’ll be an integral team member in both developing and delivering our approach to safety evaluations of Google DeepMind’s most groundbreaking models.
You will work with teams at Google DeepMind, internal and external partners, to ensure that our work is conducted in line with responsibility and safety best practices, helping Google DeepMind to progress towards its mission.
As a data scientist working in ReDI, you’ll be part of a team working on safety evaluations, using your expertise to help gather specialised data for training and evaluating our models across numerous modalities to deliver new evaluations and refine existing ones. In this role, you’ll work in collaboration with other members of this critical team, responding to needs of the business in a timely manner and prioritising projects accordingly.
Key responsibilities
Contributing to the design and development of new evaluations, particularly focussing on content policy coverage of sensitive content (such as child safety) where dataset development, rater quality and pattern analysis are primary needs
Proactively engaging with prompt dataset curation, analysis and refinement to provide feedback for iteration with 3P vendors, and opportunities for data enhancement
Investigating the behaviour of our latest models to inform evaluation design
Investigating the accuracy and patterns in human rating of evaluation outputs
Assessing the quality and coverage of safety datasets
Contributing to developing/running quantitative analyses for evaluations
Work collaboratively alongside a team of multidisciplinary specialists to deliver on priority projects
Communicating with wider stakeholders across ReDI, GDM/ Google and third party vendors where appropriate
Supporting improvements to how evaluation findings are visualised to key stakeholders and leadership
About you
In order to set you up for success as a data scientist in the ReDI team, we look for the following skills and experience:
Expertise in analytical and statistical skills, data curation and data collection design, prompt data set curation and validation
Familiarity with sociotechnical considerations of generative AI, including content safety (such as child safety) and fairness
Ability to thrive in a fast-paced, live environment where decisions are made in a timely fashion
Demonstrated ability to work within cross-functional teams, foster collaboration, and influence outcomes
Significant experience presenting and communicating data science findings to non-data science audiences, including senior stakeholders
Strong command of Python
In addition, the following would be an advantage:
Experience of working with sensitive data, access control, and procedures for data worker wellbeing
Prior experience working with product development or in similar agile settings would be advantageous
Experience in sociotechnical research and content safety
Demonstrated prior experience designing and implementing audits or evaluations of cutting edge AI systems
Experience working with ethics and safety topics associated with AI development in a technology company such as child safety, privacy, representational harms and discrimination, misinformation, or other areas of content or model risks
About the company
Artificial Intelligence could be one of humanity’s most useful inventions. At Google DeepMind, we’re a team of scientists, engineers, machine learning experts and more, working together to advance the state of the art in artificial intelligence. We use our technologies for widespread public benefit and scientific discovery, and collaborate with others on critical challenges, ensuring safety and responsibility are the highest priority.