Google DeepMind is at the forefront of artificial intelligence research and development, working to advance AI for widespread public benefit and scientific discovery. We're seeking a Software Engineer to join our Privacy Engineering team, focusing on private ML serving infrastructure for Google DeepMind and our product partners.
The role involves working on cutting-edge privacy-preserving technologies for GenAI-based agents, addressing crucial challenges in private inference using large models like Gemini, private AI agent memory, and developing privacy-preserving agentic capabilities. You'll be part of an ambitious Private Inference project, building infrastructure and researching new privacy-preserving methods.
As a Privacy Software Engineer, you'll work with state-of-the-art technology, implementing novel protocols and APIs, working with low-level components, and designing secure distributed systems at Google scale. You'll be responsible for defining threat models, implementing workloads on Trusted Execution Environments, and building transparent and secured storage systems for LLM servers.
The ideal candidate will bring strong technical expertise in security and privacy, with experience in distributed systems and proficiency in languages like Rust and C++. You'll have the opportunity to make a significant impact by creating reusable solutions for privacy-preserving AI agents across Google while contributing to the broader tech community through documentation and knowledge sharing.
Join us in shaping the future of AI privacy and security, working with a diverse team of experts committed to ethical AI development and widespread public benefit.