We have a few questions coming in, and one of them relates to something that I believe you mentioned a couple of times during the talk, which is debugging. Where are we at in terms of getting AI to debug for us? The KD, the question asker also wants to just stand and watch. Can we get there? Yes, yes we can. I think with Chrome DevTools MCP, Playwright MCP, a lot of different browser MCPs, and especially IDEs like Cursor, like other tools starting to integrate these ideas by default. I do see us getting to a place where AI agents can debug automatically, and I see that being part of this automated quality loop, where if you are prompting, they just happen to automatically use the right tools, get the right information from the browser, from anywhere else it needs context, and can hopefully avoid humans having to debug. Nice. Although again, we all still need to be kind of aware of how this, I like the context of AI pair programmer, is working alongside of us. Awesome.
Well, let's get into the, you know, the thing on everyone's mind is, where does it stop? Does it stop? How close do you think we are to hitting the limits of what AI can do for us? I think we're actually far off from hitting the limits of AI. I think that one takeaway from some of the things I talked about today is that even our ability to measure whether AI is generating code that is sufficiently high quality, something that a senior engineer would write for a production database, a production code base, or someone in an enterprise, I feel like there's still a long way to go, and so I see models still having a little bit of runway left to get better. Plenty of work still to be done, folks. Okay, well, I guess, speaking of generating tasteful designs, we have a question that, well, I guess it contains an implied question. Did you generate images for your slides? And if so, what tool did you use? I hand drew them all. No, if you saw any images of my slides, they are generated using Gemini. We use NanoBanana for these, and if you like NanoBanana, keep an eye out for other announcements this week. Oh, stay tuned, folks. Amazing, and, of course, you know, we're dogfooding. We're using the tools to talk about the tools so that we can all get better at the tools. Amazing. All right, let's see. We have a question here around security and architecture and these other kind of higher level aspects of a code base. So what is the most common security or architectural regression in AI-generated React PRs or changes, and how would you review for it? I think there's two bits here. So for the architectural regressions, what I often see, and we've done some studies around this, is that models can generate code that looks roughly right but may not be following the team's patterns correctly, because it's not necessarily looking at the full context of your code base, all the PRs, like that history, that team history. And so I think that that is a place where we do need to apply a little bit more diligence during code review. I'm hopeful that at the tooling layer, we can leverage CI, we can leverage more bots to pay attention to that and flag it as a concern before you have people actually manually reviewing the code. The most common security issues are actually very basic things, and part of it is because we've now increased the total addressable market of who can build for the web. We have a lot of people via coding who don't have a deep technical background, and so even things like XSS issues, client-side API key leaks, all of these are the biggest common problems right now. There are many more nuanced ones, but those are the really big ones, because we've got this huge audience now building. Gotcha. All right.
Comments