Flag job

Report

Data Scientist, Responsible Development and Innovation

Location

London, UK

About the role

As a data scientist in the Responsible Development and Innovation (ReDI) team, you’ll be an integral team member in both developing and delivering our approach to safety evaluations of Google DeepMind’s most groundbreaking models. You will work with teams at Google DeepMind, internal and external partners, to ensure that our work is conducted in line with responsibility and safety best practices, helping Google DeepMind to progress towards its mission. As a data scientist working in ReDI, you’ll be part of a team working on safety evaluations, using your expertise to help gather specialised data for training and evaluating our models across numerous modalities to deliver new evaluations and refine existing ones. In this role, you’ll work in collaboration with other members of this critical team, responding to needs of the business in a timely manner and prioritising projects accordingly. Key responsibilities Contributing to the design and development of new evaluations, particularly focussing on content policy coverage of sensitive content (such as child safety) where dataset development, rater quality and pattern analysis are primary needs Proactively engaging with prompt dataset curation, analysis and refinement to provide feedback for iteration with 3P vendors, and opportunities for data enhancement Investigating the behaviour of our latest models to inform evaluation design Investigating the accuracy and patterns in human rating of evaluation outputs Assessing the quality and coverage of safety datasets Contributing to developing/running quantitative analyses for evaluations Work collaboratively alongside a team of multidisciplinary specialists to deliver on priority projects Communicating with wider stakeholders across ReDI, GDM/ Google and third party vendors where appropriate Supporting improvements to how evaluation findings are visualised to key stakeholders and leadership About you In order to set you up for success as a data scientist in the ReDI team, we look for the following skills and experience: Expertise in analytical and statistical skills, data curation and data collection design, prompt data set curation and validation Familiarity with sociotechnical considerations of generative AI, including content safety (such as child safety) and fairness Ability to thrive in a fast-paced, live environment where decisions are made in a timely fashion Demonstrated ability to work within cross-functional teams, foster collaboration, and influence outcomes Significant experience presenting and communicating data science findings to non-data science audiences, including senior stakeholders Strong command of Python In addition, the following would be an advantage: Experience of working with sensitive data, access control, and procedures for data worker wellbeing Prior experience working with product development or in similar agile settings would be advantageous Experience in sociotechnical research and content safety Demonstrated prior experience designing and implementing audits or evaluations of cutting edge AI systems Experience working with ethics and safety topics associated with AI development in a technology company such as child safety, privacy, representational harms and discrimination, misinformation, or other areas of content or model risks

About the company

Artificial Intelligence could be one of humanity’s most useful inventions. At Google DeepMind, we’re a team of scientists, engineers, machine learning experts and more, working together to advance the state of the art in artificial intelligence. We use our technologies for widespread public benefit and scientific discovery, and collaborate with others on critical challenges, ensuring safety and responsibility are the highest priority.

Skills

python
data curation
data collection design
data set curation
data validation
analytical skills
statistical skills