[Acquired] Google: the AI company (Part 1)
You can't say you understand today's AI landscape without listening to this massive (4 hours!) Acquired episode on Google, focusing on its AI roots. Over three episodes, Acquired has a little over 12 hours worth of podcast just on Google! Well worth it IMO for the greatest business in history.
Selected highlights:
[07:23] basically every single person of note in AI worked at Google with the one exception of Yann Le Cun who worked at Facebook
This is truly mind-bending to think about, especially considering that Google is (at the moment) not the first name that comes to mind when we think about AI (LLMs) today. But the real kicker comes a few minutes in when we learn that did you mean? (launched in 2001!!) and google translate (2006) are the first practical application of language models to its search business which made it exponentially more effective. About 25 years ago, Google was already running machine learning in production, at fantastic scale (about 15% of Google’s entire data center infrastructure).
In 2003 AdSense uses PHIL to "understand" the advertiser page and run the Google Ad corpus against them. Again, another massive advantage for Google core business and again, 20 years before the collective AI hype.
[17:22] Maybe 10x engineers exist? If they do, Jeff Dean has to be at least 1000x: Jeff’s work with the team gets that average sentence translation time down from 12 hours to 100 milliseconds. Then they ship it in Google Translate and it’s amazing. A 432000x improvement!
Jeff is so good that they start making up Chuck Norris facts about him. My favourite: Jeff Dean's PIN is the last 4 digits of pi 😂
[35:10] Cat paper, which allows Google to build a deep neural network that can classify video content on Youtube, ultimately leading to a recommendation system working on the actual content and not on author-supplied information like title, description or tags.
Youtube is no longer a destination site.
The same technology will also be used for copyright infringement and content filtering. In 2012! Ten years before ChatGPT.
At this point it was all still running on CPUs, but it was all about to change when AlexNet was designed to run NVIDIA GeForce GTX 580s.
[1:25:00] In 2016 Google uses DeepMind to achieve a 40% reduction in the energy required to cool data centers! Again, AI has a massive, very real impact on Google bottom line.
[1:40:22] In Spring of 2014 Google buys 130 million worth of GPUs from Nvidia. This one single order is 3.25% of Nvidia revenue at the time. And with this new fast hardware, Google starts adding AI to Google Photos, Gmail autosuggestion, AdWords predicting interests, speech recognition on Android but at the same time Google realizes that they are just going to be shipping hundreds of millions soon to be billions of dollars over to NVIDIA every year for the foreseeable future.
So they set to work on the TPU and the TPU was designed, verified, built, and deployed into data centers in 15 months. FIFTEEN MONTHS!!!
And a great trivia about this rollout: they fit the TPU into the form factor of a hard drive, so it could actually slot into the existing server racks. You just pop out a hard drive and you pop in a TPU without needing to do any physical re-architecture. Amazing and clever!
And just to give an idea of the absolute scale of the TPU deployment (emphasis mine):
today it’s estimated Google estimated has 2–3 million TPUs. For reference NVIDIA shipped—people don’t know for sure—somewhere around 4 million GPUs last year.People talk about AI chips like it’s this one horse race with NVIDIA. Google has an almost NVIDIA-scale internal thing making their own chips at this point for their own and for Google Cloud customers. The TPU is a giant deal in AI in a way that I think a lot of people don’t realize.
To be continued.
Read my other AI-related posts.