Hi, everybody. It's great to see you. My name is Carly, and today I want to talk to you about context engineering. So, you might be wondering why we need to talk about context for agents. Can we not just code and let it figure it out with its information? I'm going to explain why that's not necessarily the case. We're going to talk about each of the components of context engineering with examples to show you the kind of practices you'll be doing.
And, of course, we'll have Q&A at the end after this recording. So, if you want to ask me questions, do come and join me in the Q&A session. So, if you haven't met me before, hi, it's nice to meet you. I'm Carly Richland, and I work in the developer advocacy team at Elastic. I am collecting socials like Pokemon. I'm sure a few of you are as well. So, if I don't get to your question in Q&A and you want to come and ask me something afterwards, just scan that link there. That will take you to my link tree, and you can find me wherever works. And I'm happy to help or answer questions as they come up.
So, why do we care about the context that's being passed on LLM? Well, firstly, we need to understand what the context window is. In simple terms, the context window is the number of tokens that an LLM can process at once. And while you might be thinking, well, I'm seeing from a lot of the emerging new LLMs that they have these super big context windows, it doesn't necessarily matter what I'm sending there. I'm going to disagree and I'm going to say that that's not quite right. It's still very much possible to overflow a context window. It just gets harder with the more tokens. But also, we can actually end up doing things like poisoning our context. If we have irrelevant information in there that's not actually useful, that can influence the actions that an LLM can take and therefore impact the overall result. So, it's actually still very important to make sure that the context window represents the type of information that you want it to. And not just talking about a single agent, but also making sure when we get into multi-agent architectures that each of those has the right context that it needs not only to perform its own task but to coordinate with others. It's also partially a cost point as well if we think about it. Foundational models tend to be charging per token. So, of course, there's definitely savings to make to make sure that our agents are working within our means and also making sure that we're not blowing the entire company budget when we're building these tools. It's also that LLMs do still make things up. They still hallucinate a little bit and that's down to a few different reasons.
Comments