Every API is a Tool for Agents with Code Mode

This ad is not shown to multipass and full ticket holders
JS Nation
JSNation 2026
June 11 - 15, 2026
Amsterdam & Online
The main JavaScript conference of the year
Upcoming event
JSNation 2026
JSNation 2026
June 11 - 15, 2026. Amsterdam & Online
Learn more
Bookmark
Rate this content
Sentry
Promoted
Code breaks, fix it faster

Crashes, slowdowns, regressions in prod. Seer by Sentry unifies traces, replays, errors, profiles to find root causes fast.

At Cloudflare we have a lot of products. Our REST OpenAPI spec is over 2.3 million tokens. When teams wanted to let AI agents access their services, they did what everyone does: cherry-picked important endpoints for their product, wrote some tool definitions and shipped a an MCP Server that covered a small fraction of their API.

I think we got it all wrong.

The context limit is not an MCP problem. It's an Agent problem. Tools should probably be discovered on demand. CLIs get this for free, self-discoverable and documented by design. APIs just need a little help.

This talk will cover how Code Mode works, why Dynamic Worker Loaders are super cool and how efficient sandboxes will be the great unlock for Agents. 

This talk has been presented at Node Congress 2026, check out the latest edition of this JavaScript Conference.

FAQ

MCP, or Model Context Protocol, was introduced in November 2024. It allows tools and other primitives to be hosted on a server, enabling any agent to dynamically register and use these tools.

Tool calling involves using an LLM (Large Language Model) to decide when to call a tool. The user makes a request, the LLM processes it, and if necessary, calls the appropriate tool to execute the task.

The context limit issue refers to the challenge of fitting large amounts of data, such as Cloudflare's open API spec, into an LLM's context window, which usually cannot handle such large amounts of tokens.

CodeMode is a method where the agent executes LLM-generated code in a secure environment, allowing for efficient and flexible tool usage. It is beneficial because it is token-efficient and allows complex operations to be done in a single round trip.

Running generated code presents challenges like security risks (e.g., running untrusted code). These are addressed by using a secure sandbox environment like Cloudflare's dynamic worker loaders, which isolate and control the execution of code.

The proposed system improves efficiency by allowing the execution of multiple operations in a single tool call, reducing token usage and making the interaction with APIs more seamless and less bulky.

Dynamic worker loaders are used to run untrusted code in a secure, isolated environment, ensuring that the execution is controlled and network requests can be inspected or blocked as needed.

The Cloudflare API serves as a case study for demonstrating how large APIs can be managed and accessed efficiently using the MCP and CodeMode methods, addressing challenges like context limits and tool management.

The main purpose of APIs, as discussed in the talk, is to serve as tools for agents, allowing AI to access resources outside of their initial environment, thereby extending their functionality.

The speaker proposes several methods like CLI bash and tool search for progressive disclosure, but emphasizes a new approach called CodeMode, which allows tools to be discovered on demand rather than being pre-loaded into the context.

Matt Carey
Matt Carey
32 min
26 Mar, 2026

Comments

Sign in or register to post your comment.
Video Summary and Transcription
The Talk introduces APIs, Codemode, and MCP, highlighting the role of APIs in empowering agents and AI's access to external resources. CloudFlare's evolution with MCP involves remote tool sharing and addressing challenges like large context windows. Challenges in Cloudflare's API tool access include progressive tool disclosure and splitting MCP servers. The discussion covers tool search, project management API creation, CodeMode concept, and self-documenting tools for secure tool execution. Proposed solutions include Code Mode, SDK generation, efficient control flow, token efficiency, and secure code execution. Key topics include dynamic worker loaders, Cloudflare API integration, Visitor Counter implementation, and exploring MCP server functionality.

1. Introduction to APIs, Codemode, and MCP

Short description:

Welcome to a session introducing APIs, Codemode, and MCP. Discussing the role of APIs as tools for agents. Exploring AI's access to external resources and the evolution from function calling to tool calling for agent empowerment.

Welcome to this session on APIs, Codemode, and MCP. So I guess the thesis for this is that every API is going to be a tool for agents. And we'll talk about what that means a little bit in the future. But my name is Matt. I work on agents and MCP at Cloudflare. And welcome to this talk.

My job is really, really fun because I get to work out what does the future of this stuff look like? And I get to test it and play with it. And a lot of the times, it reminds me of this meme. So I thought I'd lighten the mood with the meme. So, if a dog wore pants, would he wear them like this or like this? I have no idea. But I guess it would be cool to experiment with one or the other. The right one looks like his legs might get cold. But you never know. But this is actually what I spend my day-to-day working on.

So, like, giving agents hands. How do we let AI, which has got so useful in the last few years, how do we let it access stuff outside of the environment that we've contained it in? How do we let it access stuff outside the glass box? This might be making, like, API calls. This might be using a computer. They're doing all of this stuff. But it basically boils down into what was called function calling and now a lot of people call tool calling, giving models tools. A user asks for something, the LLM replies and decides to call this, like, tool.

This is a tool that the user will then execute and then come back and give the LLM the result. Tool calling. And I'm sure you've seen a bunch of this stuff maybe, like, in chat GPT or whatever. But this is how, in general, we give agents hands to access stuff in the outside world. We've been doing this long enough now, a few years, that we have had some new things come up. We started with just giving agents tools inside their own applications. So, each developer would write their own tools, give it to their own agent, their own LLM in a loop, and they would execute the tools when it was needed.

2. Evolution of CloudFlare's Tool Sharing via MCP

Short description:

CloudFlare's evolution with MCP from local tool bundles to remote tool sharing. Challenges of large context windows and the need for progressive tool disclosure.

And this is pretty useful for your own application. But as CloudFlare, as a company, a producer of things like APIs and a producer of a platform, we want to let our customers use our platform via these, like, LLMs, these agents. And for that, we need to, it would be really good to publish, like, more standardized set of tools that everyone else could use. And throughout some of this process, MCP was born in November of 2024, already, which is, like, a really long time ago, of what feels like yesterday. But the idea was that you could take tools and other primitives, like, resources and prompts, but mostly people have used tools, and you could host them in a server somewhere else.

And then any agent that wanted to use those tools could dynamically register with that authentication, use the tools, and say, thank you very much, really. And from a server developer point of view, it's, like, how do we share tools with, like, agents we've never met, with people we've never met? And this is not that normal on the internet. So, there was some new flow with new auth, DCR, dynamic client registration that had to be developed a little bit more. And now we're actually moving to a thing later called SIMD, Client ID Metadata Documents, like, just, like, ways that we allow agents to access stuff for things that they have no interaction with.

And it's a different type of contract from, like, an API to a client usage, because when you make that contract, you say that the API should never change. Because I hear, like, the MCP server could change their tools whenever they want. They're dynamically registered. So, it makes them very useful. But this is not really the point of the talk. This is a little bit of a background. We went from bundle tools in each application to a lot of tools being remote, shared via MCP, Model Context Protocol. And then, very quickly, we filled context windows. You've probably seen, like, memes, like, news articles or news in general about, like, there are some MCP servers that fill huge amounts of context.

Check out more articles and videos

We constantly think of articles and videos that might spark Git people interest / skill us up or help building a stellar career

Building a Voice-Enabled AI Assistant With Javascript
JSNation 2023JSNation 2023
21 min
Building a Voice-Enabled AI Assistant With Javascript
Top Content
This Talk discusses building a voice-activated AI assistant using web APIs and JavaScript. It covers using the Web Speech API for speech recognition and the speech synthesis API for text to speech. The speaker demonstrates how to communicate with the Open AI API and handle the response. The Talk also explores enabling speech recognition and addressing the user. The speaker concludes by mentioning the possibility of creating a product out of the project and using Tauri for native desktop-like experiences.
The Ai-Assisted Developer Workflow: Build Faster and Smarter Today
JSNation US 2024JSNation US 2024
31 min
The Ai-Assisted Developer Workflow: Build Faster and Smarter Today
Top Content
AI is transforming software engineering by using agents to help with coding. Agents can autonomously complete tasks and make decisions based on data. Collaborative AI and automation are opening new possibilities in code generation. Bolt is a powerful tool for troubleshooting, bug fixing, and authentication. Code generation tools like Copilot and Cursor provide support for selecting models and codebase awareness. Cline is a useful extension for website inspection and testing. Guidelines for coding with agents include defining requirements, choosing the right model, and frequent testing. Clear and concise instructions are crucial in AI-generated code. Experienced engineers are still necessary in understanding architecture and problem-solving. Energy consumption insights and sustainability are discussed in the Talk.
The Rise of the AI Engineer
React Summit US 2023React Summit US 2023
30 min
The Rise of the AI Engineer
Top Content
Watch video: The Rise of the AI Engineer
The rise of AI engineers is driven by the demand for AI and the emergence of ML research and engineering organizations. Start-ups are leveraging AI through APIs, resulting in a time-to-market advantage. The future of AI engineering holds promising results, with a focus on AI UX and the role of AI agents. Equity in AI and the central problems of AI engineering require collective efforts to address. The day-to-day life of an AI engineer involves working on products or infrastructure and dealing with specialties and tools specific to the field.
AI and Web Development: Hype or Reality
JSNation 2023JSNation 2023
24 min
AI and Web Development: Hype or Reality
Top Content
This talk explores the use of AI in web development, including tools like GitHub Copilot and Fig for CLI commands. AI can generate boilerplate code, provide context-aware solutions, and generate dummy data. It can also assist with CSS selectors and regexes, and be integrated into applications. AI is used to enhance the podcast experience by transcribing episodes and providing JSON data. The talk also discusses formatting AI output, crafting requests, and analyzing embeddings for similarity.
The AI-Native Software Engineer
JSNation US 2025JSNation US 2025
35 min
The AI-Native Software Engineer
Software engineering is evolving with AI and VIBE coding reshaping work, emphasizing collaboration and embracing AI. The future roadmap includes transitioning from augmented to AI-first and eventually AI-native developer experiences. AI integration in coding practices shapes a collaborative future, with tools evolving for startups and enterprises. AI tools aid in design, coding, and testing, offering varied assistance. Context relevance, spec-driven development, human review, and AI implementation challenges are key focus areas. AI boosts productivity but faces verification challenges, necessitating human oversight. The impact of AI on code reviews, talent development, and problem-solving evolution in coding practices is significant.
Web Apps of the Future With Web AI
JSNation 2024JSNation 2024
32 min
Web Apps of the Future With Web AI
Web AI in JavaScript allows for running machine learning models client-side in a web browser, offering advantages such as privacy, offline capabilities, low latency, and cost savings. Various AI models can be used for tasks like background blur, text toxicity detection, 3D data extraction, face mesh recognition, hand tracking, pose detection, and body segmentation. JavaScript libraries like MediaPipe LLM inference API and Visual Blocks facilitate the use of AI models. Web AI is in its early stages but has the potential to revolutionize web experiences and improve accessibility.

Workshops on related topic

AI on Demand: Serverless AI
DevOps.js Conf 2024DevOps.js Conf 2024
163 min
AI on Demand: Serverless AI
Top Content
Featured WorkshopFree
Nathan Disidore
Nathan Disidore
In this workshop, we discuss the merits of serverless architecture and how it can be applied to the AI space. We'll explore options around building serverless RAG applications for a more lambda-esque approach to AI. Next, we'll get hands on and build a sample CRUD app that allows you to store information and query it using an LLM with Workers AI, Vectorize, D1, and Cloudflare Workers.
AI for React Developers
React Advanced 2024React Advanced 2024
142 min
AI for React Developers
Top Content
Featured Workshop
Eve Porcello
Eve Porcello
Knowledge of AI tooling is critical for future-proofing the careers of React developers, and the Vercel suite of AI tools is an approachable on-ramp. In this course, we’ll take a closer look at the Vercel AI SDK and how this can help React developers build streaming interfaces with JavaScript and Next.js. We’ll also incorporate additional 3rd party APIs to build and deploy a music visualization app.
Topics:- Creating a React Project with Next.js- Choosing a LLM- Customizing Streaming Interfaces- Building Routes- Creating and Generating Components - Using Hooks (useChat, useCompletion, useActions, etc)
Building Full Stack Apps With Cursor
JSNation 2025JSNation 2025
46 min
Building Full Stack Apps With Cursor
Featured Workshop
Mike Mikula
Mike Mikula
In this workshop I’ll cover a repeatable process on how to spin up full stack apps in Cursor.  Expect to understand techniques such as using GPT to create product requirements, database schemas, roadmaps and using those in notes to generate checklists to guide app development.  We will dive further in on how to fix hallucinations/ errors that occur, useful prompts to make your app look and feel modern, approaches to get every layer wired up and more!  By the end expect to be able to run your own AI generated full stack app on your machine!
Please, find the FAQ here
Vibe coding with Cline
JSNation 2025JSNation 2025
64 min
Vibe coding with Cline
Featured Workshop
Nik Pash
Nik Pash
The way we write code is fundamentally changing. Instead of getting stuck in nested loops and implementation details, imagine focusing purely on architecture and creative problem-solving while your AI pair programmer handles the execution. In this hands-on workshop, I'll show you how to leverage Cline (an autonomous coding agent that recently hit 1M VS Code downloads) to dramatically accelerate your development workflow through a practice we call "vibe coding" - where humans focus on high-level thinking and AI handles the implementation.You'll discover:The fundamental principles of "vibe coding" and how it differs from traditional developmentHow to architect solutions at a high level and have AI implement them accuratelyLive demo: Building a production-grade caching system in Go that saved us $500/weekTechniques for using AI to understand complex codebases in minutes instead of hoursBest practices for prompting AI agents to get exactly the code you wantCommon pitfalls to avoid when working with AI coding assistantsStrategies for using AI to accelerate learning and reduce dependency on senior engineersHow to effectively combine human creativity with AI implementation capabilitiesWhether you're a junior developer looking to accelerate your learning or a senior engineer wanting to optimize your workflow, you'll leave this workshop with practical experience in AI-assisted development that you can immediately apply to your projects. Through live coding demos and hands-on exercises, you'll learn how to leverage Cline to write better code faster while focusing on what matters - solving real problems.
The React Developer's Guide to AI Engineering
React Summit US 2025React Summit US 2025
96 min
The React Developer's Guide to AI Engineering
Featured WorkshopFree
Niall Maher
Niall Maher
A comprehensive workshop designed specifically for React developers ready to become AI engineers. Learn how your existing React skills—component thinking, state management, effect handling, and performance optimization—directly translate to building sophisticated AI applications. We'll cover the full stack: AI API integration, streaming responses, error handling, state persistence with Supabase, and deployment with Vercel.Skills Translation:- Component lifecycle → AI conversation lifecycle- State management → AI context and memory management- Effect handling → AI response streaming and side effects- Performance optimization → AI caching and request optimization- Testing patterns → AI interaction testing strategiesWhat you'll build: A complete AI-powered project management tool showcasing enterprise-level AI integration patterns.
Build LLM agents in TypeScript with Mastra and Vercel AI SDK
React Advanced 2025React Advanced 2025
145 min
Build LLM agents in TypeScript with Mastra and Vercel AI SDK
Featured WorkshopFree
Eric Burel
Eric Burel
LLMs are not just fancy search engines: they lay the ground for building autonomous and intelligent pieces of software, aka agents.
Companies are investing massively in generative AI infrastructures. To get their money's worth, they need developers that can make the best out of an LLM, and that could be you.
Discover the TypeScript stack for LLM-based development in this 3 hours workshop. Connect to your favorite model with the Vercel AI SDK and turn lines of code into AI agents with Mastra.ai.