Hongil Yoon Profile Picture

Hongil Yoon

Technical Lead & Manager, Google | Visiting Associate Professor, SNU
Mountain View, CA, USA  •  hongilyoon@google.com  •  LinkedIn

About Me

I am a Technical Lead & Manager at Google and a Visiting Associate Professor at the AI Institute of Seoul National University (AIIS). My research bridges the gap between computer architecture and machine learning, with a specific focus on hardware-software co-design for efficient AI computing. At Google, I lead engineering teams developing next-generation TPU architectures and system optimizations for cloud-scale training and serving. Previously, I architected the EdgeTPU and memory subsystems for the Google Tensor SoC, powering the Pixel series.

Currently, my academic research explores efficient ML training & inference infrastructure and 3D perception systems, tackling the computational and memory challenges of Large-Scale 3D Gaussian Splatting and point cloud processing via novel host-offloading and pruning techniques. I received my Ph.D. in Computer Sciences from the University of Wisconsin–Madison. I was advised by Professor Gurindar S. Sohi.

Work Experience

Google Inc.
Software Engineer, Technical Lead & Manager Apr 2025 – Present
  • TPU hardware-software co-design and system architecture exploration for cloud-scale ML deployments.
  • Leading specialized engineering teams focused on optimizing large-scale pre-training and serving for next-generation AI architectures.
ML Accelerator Architect, Technical Lead Nov 2018 – Apr 2025
  • Technical Lead for Generative AI Compute & Memory Architecture for the Tensor SoC, defining long-term strategic directions.
  • Spearheaded development of EdgeTPU and advanced memory systems for the Google Pixel ecosystem (Pixel 7 through Pixel 11).
SoC Performance Architect Apr 2017 – Nov 2018
  • Conducted architectural modeling and performance analysis for future-generation Tensor SoC for Pixel 6.
Seoul National University
Visiting Associate Professor (June 2024–Present), Artificial Intelligence Institute Seoul National University (AIIS), South Korea
  • Awarded Google Gift (2024/2025) for Seoul National University ARC LAB.
  • GS-Scale: Unlocking Large-Scale 3D Gaussian Splatting Training via Host Offloading, ASPLOS 2026.
  • FastPoint: Accelerating 3D Point Cloud Model Inference via Sample Point Distance Prediction, ICCV 2025.
  • Frugal 3D Point Cloud Model Training via Sample Recycling and Fused Aggregation, ECCV 2024.
University of Wisconsin–Madison
Research Assistant Jun 2010 – Mar 2017
  • Developed novel virtual caching and memory hierarchy techniques to significantly reduce address translation overheads in modern processors.

Education

University of Wisconsin–Madison May 2017
Ph.D. in Computer Sciences
Dissertation: Reducing Address Translation Overheads with Virtual Caching. [Dissertation and Defense Talk]
Committees: Gurindar S. Sohi (Chair), Mark D. Hill, Mikko H. Lipasti, Karthikeyan Sankaralingam, and David A. Wood.
University of Wisconsin–Madison May 2012
M.S. in Computer Sciences
Korea University February 2007
B.E. in Computer Sciences and Engineering

Publications

Encompassing diverse areas within computer architecture and machine learning. My early work focused on enhancing caching mechanisms and memory systems. I later expanded this scope to mobile computer architecture, addressing the unique constraints of on-device ML. Currently, my research focuses on the hardware-software co-design of efficient 3D perception systems, specifically optimizing 3D point cloud processing and Large-Scale Gaussian Splatting through novel ML training techniques and host-offloading strategies.

*Corresponding authors

Patents

Professional Activities and Service