Google's software engineers develop the next-generation technologies that change how billions of users connect, explore, and interact with information and one another. Our products need to handle information at massive scale, and extend well beyond web search. We're looking for engineers who bring fresh ideas from all areas, including information retrieval, distributed computing, large-scale system design, networking and data storage, security, artificial intelligence, natural language processing, UI design and mobile; the list goes on and is growing every day. As a software engineer, you will work on a specific project critical to Google's needs with opportunities to switch teams and projects as you and our fast-paced business grow and evolve. We need our engineers to be versatile, display leadership qualities and be enthusiastic to take on new problems across the full-stack as we continue to push technology forward. Cloud ML Compute Services (CMCS) team is within the Cloud organization that has been chartered to build x-Google alignment towards a unified central infrastructure to host all of Google's ML needs, including internal and external use cases. The CMCS Inference team is part of the CMCS team and focuses on the inference workloads and the serving infrastructure. In this role, you will be optimizing machine learning models for large scale inference workloads and will also have experience in different large scale Machine Learning (ML) optimizations techniques for improving latency and throughput. You will have experience with accelerators (TPUs or GPUs), or HPC.
Stand Out From the Crowd
Upload your resume and get instant feedback on how well it matches this job.
Job Type
Full-time
Career Level
Mid Level
Industry
Web Search Portals, Libraries, Archives, and Other Information Services