NodeJS & AI: Building Smarter Applications

This ad is not shown to multipass and full ticket holders
JSNation US
JSNation US 2025
November 17 - 20, 2025
New York, US & Online
See JS stars in the US biggest planetarium
Learn More
In partnership with Focus Reactive
Upcoming event
JSNation US 2025
JSNation US 2025
November 17 - 20, 2025. New York, US & Online
Learn more
Bookmark
Rate this content

Unlock the potential of AI in Node.js applications through practical, real-world implementations. This session demonstrates how to integrate both open-source and closed AI models to build intelligent features like semantic search and autonomous agents. Attendees will learn production-ready patterns for RAG systems, vector embeddings, and AI agents while understanding the core architectural decisions behind each implementation. Perfect for developers ready to move beyond basic API calls to building sophisticated AI-enhanced applications.

This talk has been presented at Node Congress 2025, check out the latest edition of this JavaScript Conference.

FAQ

A model in AI is a specialized program trained on massive amounts of data to recognize patterns and make predictions or decisions. These models perform specific tasks, such as understanding text, generating content, or classifying images.

Large language models are designed to understand and generate human language, and they are based on transformer architecture. The foundational paper for this architecture is 'Attention is All You Need.'

Closed models, like ChatGPT or Gemini, are accessed via API and come with a cost, whereas open-source models can be downloaded and run locally or hosted on cloud infrastructure. Examples of open-source models include Lama and DeepSeq.

RAC, which stands for Retrieve of Method Generation, helps reduce hallucination in AI models by providing relevant information to improve accuracy and enable the use of private data for more powerful models.

Embedding transforms data, particularly text, into a list of numbers to capture its semantic meaning. This positions the text in a high-dimensional space where similar meanings are closer together, facilitating retrieval tasks like similarity search.

Vectors represent the semantic meaning of text and are stored in vector databases. They allow for similarity searches to retrieve information based on how close and semantically similar the meanings are.

The retrieval techniques mentioned include keyword or term-based, semantic, and hybrid retrieval. In the example, semantic retrieval is used to match user queries with relevant information.

An agent in AI is a system that works to achieve specific goals by planning, reasoning, and using tools. Unlike simple prompt-response models, agents plan their approach, maintain context, and adapt to fulfill tasks rather than just provide static responses.

The React pattern involves defining functions (tools), setting a mission or goal, reasoning through the mission, executing actions with appropriate tools, observing the results, and synthesizing a response based on the gathered context.

Aileen Villanueva
Aileen Villanueva
19 min
17 Apr, 2025

Comments

Sign in or register to post your comment.
  • Christos kuznos
    Christos kuznos
    None
    are they more than the 2 types of models, closed and open source?
Video Summary and Transcription
Today's Talk explored combining Node.js and artificial intelligence to build smart applications. Two types of models, closed and open source, were discussed, with closed models accessed via an API and open source models hosted locally or on the cloud. The use of RAC (Retrieve of Method Generation) to enhance models was explained, along with the process of obtaining data and improving retrieval performance. Chunking, embedding, and vector databases were introduced as techniques for organizing and transforming data. The Talk also covered data retrieval using semantic search and the generation of human-readable answers using models. The concept of enhancing models with agents was discussed, focusing on how agents break down complex goals and utilize external functions. The use of tools and the React pattern in implementing agents was highlighted. Lastly, the Talk touched on implementing agent patterns and enhancing LLMs, with demos and code available for further exploration.

1. Introduction to Node.js and AI

Short description:

Today we'll explore combining Node.js and artificial intelligence to build smart applications. Models are specialized programs trained on data to recognize patterns and make predictions. We'll be using large language models that excel at understanding and generating human language. There are two types of models: closed and open source. Closed models like chat GPT and Gemini are accessed via an API. Open source models like Lama and Gemma can be hosted locally or on the cloud. While these models are powerful, they have limitations. That's where new techniques come into play.

Hi all, I'm excited to talk today about how we can combine Node.js and one part of artificial intelligence to build applications that aren't just functional, but smart. There has been an increasing importance of artificial intelligence in software, and I want to encourage you TypeScript and JavaScript developers to learn and apply these technologies.

My name is Aileen. I'm currently working as a software developer at Orama, and I've been doing software for 10 years now. I organize technology meetups at my city, and I'm really passionate about the intersection of AI and software.

But before we dive in, I want us to align with some concepts that we'll be talking about. And let's start with defining what is a model. And I want you to think of a model as a specialized program that has been trained on massive amounts of data to recognize patterns and make predictions or decisions. These models have been trained to do specific tasks, like understanding text, generating content or classifying images. And we're going to interact with them by providing these models an input. It would use all of these learned patterns and generate an output. This output could be an inference, a generation or a prediction.

Today we're going to be using large language models. And a large language model is a model that is good at understanding and generating human language. They're based on transformer architecture. There's a paper called Attention is All You Need, and I really recommend it to read it if you really want to understand how they work underneath it. They're trained on large amounts of data, and they work for many tasks. So they can do summarizations. They can do question answering or translation. Today we're going to talk about two types of models, which is the closed models. And these models are models like chat GPT, like Cloud, From Anthropic, like Gemini, and some of Mistral models. These models are accessed via an API, and they come with a cost. You cannot download it, so you basically rely on these providers to use these models. The other type of models are open source models. And these models, you can find them hosted in Hug and Face. You can download them and run them on your local machine using Olama. You could also download them and host them on any cloud infrastructure, and these models are models like Lama, like DeepSeq, or like Gemma from Google. But as you can see, these models are really powerful, and we can definitely make them more powerful. Let's remember that these models have been trained with huge amounts of data, and it's no question that these models are not good at math or very specific needs or niche of an industry. So that's when a new technique comes into play.

2. Using RAC to Enhance Models

Short description:

RAC (Retrieve of Method Generation) reduces hallucination by providing accurate information to models. Private data use and RAC can make models more powerful. However, models may be outdated or lack knowledge. For example, when asked about the speakers of Node Congress 2025, a model's knowledge was cut off in October 2024. To address this, we'll walk through creating and enhancing the model with RAC. The first step is obtaining data, which can be in various formats like PDFs, JSON, or crawled from websites. Next, we use splitting to make it easier to find relevant information and improve retrieval performance. The LimeChain library and recursive characters text splitter function can be used for this purpose.

And that technique is RAC. RAC stands for Retrieve of Method Generation. And RAC definitely helps out on things like hallucination. While it does not eliminate hallucination, it could definitely reduce it. Because if a model doesn't know something, it might have made up an answer that's not accurate. So with RAC, if we provide that information, that model can answer correctly. Or we can enable private data use. So definitely, if they were not trained with data that you have available only for you, it would definitely help make your model more powerful.

But also, let's remember that maybe all of these models are not trained daily. So there's knowledge that they don't have, or they might be outdated. And there is a specific use case that I want to show you. So I asked Claude who are going to be the speakers of Node Congress 2025. And as you can see in the answer, it mentions that the knowledge of that model at this current date was cut off on October 2024. So it doesn't know who are going to be the speakers of this conference. And so for today's exercise, I'm going to go and walk you through how to create and enhance this model with RAC to actually answer this question. We're going to follow a couple of steps. And our first step is data. So there's many ways that you can encounter data for this improvement of our models with RAC. And that could be a PDF, that could be a JSON, a CSV file, a crawler, a DocuSign site, or a database. In this case, I'm using a crawler to go into all of the Node Congress web page. And I'm just adding some markdown files that I'll later have available for them to read. So this is an example of how the markdown looks based on the crawler that I did for this web page.

The next step is the split. So large documents can be overwhelming, and our AI model can make it hard for you to find relevant information across all of this data. So with splitting, we're going to cut it and maybe make it easier to look for something like in a book. So you can look for information in a chapter rather than reading the whole book. So split is a good technique that is used to improve retrieval performance. And also, it's very good at representing context. In here, I'm using a library called LimeChain, and I'm using the function recursive characters text splitter. This is also something that you will need to decide based on your use case.

3. Chunking, Embedding, and Vector Databases

Short description:

To chunk information, we can divide by sentences, paragraphs, or characters. Chunk overlaps are crucial to maintain context. Embedding transforms text into a list of numbers, positioning similar meanings closer together in a high dimensional space. OpenAI Ada is an embedding model used to transform chunks of data. Vector databases store the content and corresponding vectors, enabling retrieval through similarity search.

You could divide by sentences, you could divide by paragraphs, you could divide by characters and how many sentences, how many paragraphs, and how many characters. There's some things that would be very specific of your use case that you will really need to try and test which one works for you. In this case, I'm using a chunk size and characters of 15,000 and a chunk overlap of 150.

Chunk overlaps are crucial and it basically ensures that we don't lose context at each chunk. So as we are going to chunk all of this information, it's important to have a small preview of what's next and also from the next chapter or if we're talking about an example of books. So overlaps are very important to have. So what we're doing is basically going through each of this information, characters, chunk them, and then now that we have this chunk of data, what's next? We're going to go through something called embedding.

So embedding is a technique that transform data, in this case text, to capture the semantic meaning of it. So each embedding is essentially a list of numbers, typically between 700 and 1,000, and it positions the text in a high dimensional space where similar meanings are closer together. As I mentioned, it's 700 and 1,000 dimensions. If we want to do it or see it in a very basic example, we can look at this image from SuperBase. So this is a two dimension and how it's going to be mapping things. So if I have the sentence, the cat chases the mouse, it's very similar to the next sentence. So they're going to be close together. But another sentence that has nothing to do with it, it would be far. And this is just in a two dimension. So we're going to rely on embedding models to transform these chunks of data and put it in a dimension space so that later we can retrieve that information based on how close and semantically these meanings are.

So as I mentioned, we're going to use an embedding model. In this case, I'm using OpenAI Ada. And I'm going to go from each of these chunks and make it an embedding. So the chunks that I did for 15,000, these are going to transform into an embedding. And it's basically vectors, a representational number of where they're put in a dimension. And now that I have this data, I'm going to need a place to find it later on. And this is where vector databases come into play. I will need to start these data in this special type of databases, vector databases. And it would look something like this. Remember, I have my 15,000 character content. So this is basically the content. And in the other side, I have all the vectors that have the meaning of that content. This is going to be used to later retrieve this information with something called similarity search.

4. Data Retrieval and Semantic Search

Short description:

To retrieve data, there are different types of retrievals: keyword or term-based, semantic, and hybrid. We'll use semantic retrieval for this case. The process involves transforming the user query into an embedding and conducting a similarity search to find similar information. We'll follow step-by-step instructions, using the same model for embedding and retrieval.

Awesome. So we have our data. We have split it. And we have stored it. Now we need to retrieve that data. How are we going to find that information later? There are many types of retrievals that we can use. This can be keyword or term-based, which is basically looking for that exact query or word or sentence. There's also semantic, which is most likely looking to similar things based on the meaning. And there is hybrid. So basically, it's a combination of the two. This is also something that you will need to rely on based on your use case. On some places, it might be great to use one or the other. And this is something that you might want to explore based on what gives you the best results for your user, which is the goal.

In this case, we're going to use semantic. And so in order to retrieve, we're going to follow these steps. First, we will need the user query. In this case, we need that question. So the question is, who are the speakers? So that would be our query. In order to make a similarity search, we need to transform that query into an embedding 2. That's the only way we could actually find something similar to it. And this would find anything that's similar and get it back. So step by step. We have an endpoint API search. We're going to get the embedding of the query, which is anything that the user asks. And then we would find anything that matches that embedding in a similarity search. And finally, we're going to generate a response.

So let's go step by step. Again, I'm using the same model that I used before, in order that it can actually make sense. It's not a good idea to use different embeddings for different for storing and for querying. So I'm using the same model.

5. Similarity Search and Human Response

Short description:

To retrieve the results of the similarity search, I use the match conference docs function in super base. I set a threshold of 0.5 for similarity and limit the number of results to eight. After obtaining the results, I generate a human-readable answer using the Gemini model. By providing the context and query, the model produces information about the speakers at node Congress, including their names, companies, and talk titles.

Now that I have my embedding, I would do a similarity search. This is going to go for a function called match conference docs in super base. With the embedding that I have transformed. And I'm going to get... I want to match threshold of 0.5, which is basically 50% similarity. Anything that's 50% above, I want it back. But only give me eight. If I have 10%, I only want eight back. And this is something that also, it's something that you can change. This is not the right answer for all of the use cases. This is the one that I use for mine. So once I have this, what's next?

I need to generate. Remember large language models are good at generating human language. So I can not just give back the results. I need to give a human response. So in this case, what I'm using is a Gemini model, and I'm going to give the context. The context is going to be everything that I retrieve from my vector database. And the query. So I'm saying to Gemini model, I need information about the following questions regarding node Congress, which is the query. And then here is the information that is relevant, which is the context. And this model will generate a human readable answer.

So let's see it in action. So this is the web page, and I'm asking who's speaking at node Congress. Great. So if you see, I'm getting the name, the companies, and even the title of the conference talk of these speakers. And this is something that we made more powerful. So we are relying on a model, a large language model, and we were able to answer this question because we provided this answer.

6. Enhancing Models with Agents

Short description:

Agents are a way to enhance models by working to achieve specific goals. Unlike simple question-answering systems, agents plan their approach, use tools, and adapt when faced with obstacles. They break complex goals into logical steps, think step by step through problems, maintain context, and utilize external functions. To implement agents, we start by defining functions.

So if you see, I'm getting the name, the companies, and even the title of the conference talk of these speakers. And this is something that we made more powerful. So we are relying on a model, a large language model, and we were able to answer this question because we provided this answer. We could provide an answer to the user and we didn't hallucinate on creating fake information.

Great. So what's next? There are many, there's another way that we could definitely enhance these models. And this is by agents. But what is an agent? And an agent is a system that don't just respond to prompts. So it's not how we were doing it. As I mentioned, Cloud asking questions. They actually work to achieve those goals. So if a user have a question that agent is gonna fulfill that goal that was made to do. So think it's very different from what we did on Cloud asking. It is very similar as if you have used coding assistant or any planning for a trip or something like that. So it would give you information, but it would try to accomplish a task. And what makes them special is because they plan their approach, they use tools when needed, and they adapt when they hit a roadblock. So instead of returning an answer that they don't know, they would try to look and adapt and use something to fulfill the mission or the goal.

Planning. Agents break complex goals into logical steps. If you ask for a competitive analysis of five companies, an agent will plan how to research each of one, what aspects to compare and how to structure the final report. Reasoning. This is where a chain of thought comes into play. So agents think step by step through problems, just like we do into solving something that is complex step by step. Agents maintain context across interactions. So they remember what they've learned and build on previous work. This is very common on coding assistants. And remember, tools. Tools is one of the most important things, and this is crucial because they can call external functions like searching the web or querying a database or making calculations. So these basically extend a capability that goes beyond the training data of the model.

So how does it look like in code? So first off, we're going to define functions.

7. Using Tools and the React Pattern

Short description:

The agent uses tools to search for information and has functions available for retrieving event information and hotel data. The agent's mission is to assist users with their travel queries. The React pattern is used for reasoning, with the model analyzing the user request and executing appropriate functions. The agent collects data through observations and synthesizes a response with context. As an AI trip planner, the agent can find hotels based on user specified criteria.

This is basically the tools that the agent can use to look for information. So we're going to do a trip, a planning trip agent. So I have these functions available. The LLM might not be trained with the whole list of hotels or events. So we're going to do something to search for this information. So I have these functions.

One, retrieve event information based on query, option filters, or finding hotel API. So these are two functions that the agent has available. We're going to need a mission. So here we define what is the mission of this agent. What is the goal? So you are an AI trip planning assistant. Help the user with their travel query. So this is the mission that the agent needs. It's very important to define that.

Next, it's the React pattern. Basically it's the reasoning. The model analyzes the user request through the mission. So I have a mission, but I also need the model to analyze it. We have the action. What is the next step? The model calls the appropriate functions if needed. And then execute the function to get that data. Once I have that, I do an observation, which is collect the data from the function calls. And then finally, synthesize into response with a context. So I'm going to rely on what are the steps that the agent needs to do. And then just put it all together to give a response.

Let's see how this works. This is a AI trip planner. And I'm finding hotels from New York for the next month. So this is information that I have, but it's also asking for more context. What are the days? What are the budget? Where are you going to be staying? What are the priorities? So this can definitely give a more specific answer based on the user needs.

8. Implementing Agent Patterns and Enhancing LLMs

Short description:

The agent uses information from the hotel API and can also search for available rooms based on guest requirements. Implementing the agent's patterns and needs helps fulfill its mission and enhances LLMs. Feel free to connect with me for any questions or further discussions. Demos will be available with the code. Thank you for joining!

Right now, it's already given information from hotels that it's using from the hotel API. And it's also asking for more information, so it can actually go and look for rooms available based on the guests that are going to be needed for it. So this is how it works.

We have these patterns and we have these needs in order to make an agent work and fulfill the mission that it has. There are many common use cases, as we mentioned previously. And this is a very powerful tool that would also enhance our LLMs.

I hope you find this talk amazing and you learned a lot. And please connect with me if you have any questions or you would like to further ask about it. These demos are going to be available with the code. And yeah. Thank you so much for joining.

Check out more articles and videos

We constantly think of articles and videos that might spark Git people interest / skill us up or help building a stellar career

It's a Jungle Out There: What's Really Going on Inside Your Node_Modules Folder
Node Congress 2022Node Congress 2022
26 min
It's a Jungle Out There: What's Really Going on Inside Your Node_Modules Folder
Top Content
The talk discusses the importance of supply chain security in the open source ecosystem, highlighting the risks of relying on open source code without proper code review. It explores the trend of supply chain attacks and the need for a new approach to detect and block malicious dependencies. The talk also introduces Socket, a tool that assesses the security of packages and provides automation and analysis to protect against malware and supply chain attacks. It emphasizes the need to prioritize security in software development and offers insights into potential solutions such as realms and Deno's command line flags.
Building a Voice-Enabled AI Assistant With Javascript
JSNation 2023JSNation 2023
21 min
Building a Voice-Enabled AI Assistant With Javascript
Top Content
This Talk discusses building a voice-activated AI assistant using web APIs and JavaScript. It covers using the Web Speech API for speech recognition and the speech synthesis API for text to speech. The speaker demonstrates how to communicate with the Open AI API and handle the response. The Talk also explores enabling speech recognition and addressing the user. The speaker concludes by mentioning the possibility of creating a product out of the project and using Tauri for native desktop-like experiences.
The Ai-Assisted Developer Workflow: Build Faster and Smarter Today
JSNation US 2024JSNation US 2024
31 min
The Ai-Assisted Developer Workflow: Build Faster and Smarter Today
Top Content
AI is transforming software engineering by using agents to help with coding. Agents can autonomously complete tasks and make decisions based on data. Collaborative AI and automation are opening new possibilities in code generation. Bolt is a powerful tool for troubleshooting, bug fixing, and authentication. Code generation tools like Copilot and Cursor provide support for selecting models and codebase awareness. Cline is a useful extension for website inspection and testing. Guidelines for coding with agents include defining requirements, choosing the right model, and frequent testing. Clear and concise instructions are crucial in AI-generated code. Experienced engineers are still necessary in understanding architecture and problem-solving. Energy consumption insights and sustainability are discussed in the Talk.
AI and Web Development: Hype or Reality
JSNation 2023JSNation 2023
24 min
AI and Web Development: Hype or Reality
Top Content
This talk explores the use of AI in web development, including tools like GitHub Copilot and Fig for CLI commands. AI can generate boilerplate code, provide context-aware solutions, and generate dummy data. It can also assist with CSS selectors and regexes, and be integrated into applications. AI is used to enhance the podcast experience by transcribing episodes and providing JSON data. The talk also discusses formatting AI output, crafting requests, and analyzing embeddings for similarity.
The Rise of the AI Engineer
React Summit US 2023React Summit US 2023
30 min
The Rise of the AI Engineer
Top Content
Watch video: The Rise of the AI Engineer
The rise of AI engineers is driven by the demand for AI and the emergence of ML research and engineering organizations. Start-ups are leveraging AI through APIs, resulting in a time-to-market advantage. The future of AI engineering holds promising results, with a focus on AI UX and the role of AI agents. Equity in AI and the central problems of AI engineering require collective efforts to address. The day-to-day life of an AI engineer involves working on products or infrastructure and dealing with specialties and tools specific to the field.
ESM Loaders: Enhancing Module Loading in Node.js
JSNation 2023JSNation 2023
22 min
ESM Loaders: Enhancing Module Loading in Node.js
Top Content
ESM Loaders enhance module loading in Node.js by resolving URLs and reading files from the disk. Module loaders can override modules and change how they are found. Enhancing the loading phase involves loading directly from HTTP and loading TypeScript code without building it. The loader in the module URL handles URL resolution and uses fetch to fetch the source code. Loaders can be chained together to load from different sources, transform source code, and resolve URLs differently. The future of module loading enhancements is promising and simple to use.

Workshops on related topic

AI on Demand: Serverless AI
DevOps.js Conf 2024DevOps.js Conf 2024
163 min
AI on Demand: Serverless AI
Top Content
Featured WorkshopFree
Nathan Disidore
Nathan Disidore
In this workshop, we discuss the merits of serverless architecture and how it can be applied to the AI space. We'll explore options around building serverless RAG applications for a more lambda-esque approach to AI. Next, we'll get hands on and build a sample CRUD app that allows you to store information and query it using an LLM with Workers AI, Vectorize, D1, and Cloudflare Workers.
AI for React Developers
React Advanced 2024React Advanced 2024
142 min
AI for React Developers
Top Content
Featured Workshop
Eve Porcello
Eve Porcello
Knowledge of AI tooling is critical for future-proofing the careers of React developers, and the Vercel suite of AI tools is an approachable on-ramp. In this course, we’ll take a closer look at the Vercel AI SDK and how this can help React developers build streaming interfaces with JavaScript and Next.js. We’ll also incorporate additional 3rd party APIs to build and deploy a music visualization app.
Topics:- Creating a React Project with Next.js- Choosing a LLM- Customizing Streaming Interfaces- Building Routes- Creating and Generating Components - Using Hooks (useChat, useCompletion, useActions, etc)
Vibe coding with Cline
JSNation 2025JSNation 2025
64 min
Vibe coding with Cline
Featured Workshop
Nik Pash
Nik Pash
The way we write code is fundamentally changing. Instead of getting stuck in nested loops and implementation details, imagine focusing purely on architecture and creative problem-solving while your AI pair programmer handles the execution. In this hands-on workshop, I'll show you how to leverage Cline (an autonomous coding agent that recently hit 1M VS Code downloads) to dramatically accelerate your development workflow through a practice we call "vibe coding" - where humans focus on high-level thinking and AI handles the implementation.You'll discover:The fundamental principles of "vibe coding" and how it differs from traditional developmentHow to architect solutions at a high level and have AI implement them accuratelyLive demo: Building a production-grade caching system in Go that saved us $500/weekTechniques for using AI to understand complex codebases in minutes instead of hoursBest practices for prompting AI agents to get exactly the code you wantCommon pitfalls to avoid when working with AI coding assistantsStrategies for using AI to accelerate learning and reduce dependency on senior engineersHow to effectively combine human creativity with AI implementation capabilitiesWhether you're a junior developer looking to accelerate your learning or a senior engineer wanting to optimize your workflow, you'll leave this workshop with practical experience in AI-assisted development that you can immediately apply to your projects. Through live coding demos and hands-on exercises, you'll learn how to leverage Cline to write better code faster while focusing on what matters - solving real problems.
Building Full Stack Apps With Cursor
JSNation 2025JSNation 2025
46 min
Building Full Stack Apps With Cursor
Featured Workshop
Mike Mikula
Mike Mikula
In this workshop I’ll cover a repeatable process on how to spin up full stack apps in Cursor.  Expect to understand techniques such as using GPT to create product requirements, database schemas, roadmaps and using those in notes to generate checklists to guide app development.  We will dive further in on how to fix hallucinations/ errors that occur, useful prompts to make your app look and feel modern, approaches to get every layer wired up and more!  By the end expect to be able to run your own AI generated full stack app on your machine!
Please, find the FAQ here
Free webinar: Building Full Stack Apps With Cursor
Productivity Conf for Devs and Tech LeadersProductivity Conf for Devs and Tech Leaders
71 min
Free webinar: Building Full Stack Apps With Cursor
Top Content
WorkshopFree
Mike Mikula
Mike Mikula
In this webinar I’ll cover a repeatable process on how to spin up full stack apps in Cursor.  Expect to understand techniques such as using GPT to create product requirements, database schemas, roadmaps and using those in notes to generate checklists to guide app development.  We will dive further in on how to fix hallucinations/ errors that occur, useful prompts to make your app look and feel modern, approaches to get every layer wired up and more!  By the end expect to be able to run your own ai generated full stack app on your machine!
Working With OpenAI and Prompt Engineering for React Developers
React Advanced 2023React Advanced 2023
98 min
Working With OpenAI and Prompt Engineering for React Developers
Top Content
Workshop
Richard Moss
Richard Moss
In this workshop we'll take a tour of applied AI from the perspective of front end developers, zooming in on the emerging best practices when it comes to working with LLMs to build great products. This workshop is based on learnings from working with the OpenAI API from its debut last November to build out a working MVP which became PowerModeAI (A customer facing ideation and slide creation tool).
In the workshop they'll be a mix of presentation and hands on exercises to cover topics including:
- GPT fundamentals- Pitfalls of LLMs- Prompt engineering best practices and techniques- Using the playground effectively- Installing and configuring the OpenAI SDK- Approaches to working with the API and prompt management- Implementing the API to build an AI powered customer facing application- Fine tuning and embeddings- Emerging best practice on LLMOps