Search latency budgets are measured in single-digit milliseconds
βBut to give an example, like, search, you know, I was speaking with the teams. Right? Like, they now have for sub teams, like, latency budgets, like, in the milliseconds. You'll get 50% credit if so if you ship something which, you know, shaves off three milliseconds, you earn one point five milliseconds for your latency budget, and one point five milliseconds gets passed on to the user.β
βI'm coming here as we just shipped Gemma four. And, it's a really good open source model. The frontier to Gemma four is both huge and not so huge in terms of time. Like, of, Gemma four is based on Gemini three architecture. You know, it's a very weird thing. Right? You're talking about a set of ways which can fit on a USB stick.β
Younger AI-native companies have a structural advantage over incumbents
βI think your question earlier on, like, you know, I think you were asking in the context of way more robotics, like, companies. I do think companies which are that's one advantage startups are gonna have. More AI native teams. And and, you know, you can probably get at it through your interview processes, etcetera. Whereas for us, we would have, like, retraining, transformation etcetera. And I think that that's maybe an advantage, like, the younger companies are gonna have.β
Sundar spends a dedicated hour weekly tracking compute allocation by team
βBut now it is really acutely constrained. Right? So you spend a lot more time. I at least spend a dedicated hour a week thinking about that question at a pretty granular level. So I will know by projects and by teams the compute units they are using. Right? And, you know, or or at least I have that information, and I'm looking at it and assessing it.β
Google was built for the AI moment but had to execute better
βHey. The overton window shifted. We have like, I felt like the company was built for that moment. The vertical thing, it's it's it's not an accident or something. It was a very intentful we were on the seventh version of TPUs. So to me, we were behind in terms of frontier LLM models, but we had all the capabilities internally, and we had to execute to meet the moment.β
Transformers were built to solve product problems, not just research
βTransformers was done in the context of a lot of, like, like, TPUs, transformers were all done to solve a specific product need to some extent. Right? Like, the team's thinking about how to make translation better. In the case of TPUs, how do you, pay speech rec works? We suddenly have to serve it to 2,000,000,000 people. We don't have enough chips for it.β
Memory is the most acute supply constraint in 2026
βMemory is definitely one of the most critical components now. There is no way that the leading memory companies are going to dramatically improve their capacity. So you have those constraints in the short term, but they they get more relaxed as you go out. By the way, I think it'll push a lot of innovations on. We will make these things 30 x more efficient.β
Sundar increased Waymo investment when others got pessimistic
βWaymo was a great example where I think we increased our investment two to three years ago when the rest of the world got pessimistic on it. When others, some of the people are backing off. For example, if Waymo had reached this point earlier, I think I would have invested the capital earlier. But I would have been glad to invest more capital in Waymo earlier, but we weren't at the level of maturity needed to do that.β
Google is in early stages of building data centers in space
βWe are constantly trying to take these long term projects, which when you first announce them, slightly marginally looks ridiculous. Okay. You know, like, we're in the earliest stages of thinking about data centers in space. But to your earlier discussion around constraint inspires creativity. But if you take a twenty year outlook, right, where are you going to put most of these data centers? Really hard problems to solve.β