Bring AI-Based Search to Your Web App

This ad is not shown to multipass and full ticket holders
JSNation US
JSNation US 2025
November 17 - 20, 2025
New York, US & Online
See JS stars in the US biggest planetarium
Learn More
In partnership with Focus Reactive
Upcoming event
JSNation US 2025
JSNation US 2025
November 17 - 20, 2025. New York, US & Online
Learn more
Bookmark
Rate this content

ChatGPT took the tech world by storm. Everyone talks about it, from your CTO to your hairdresser (at least my barber does). And there are many reasons why we should all be excited about it and many other AI/ML innovations.


But how do you bring them into your tech stack, your website/backend, to work with your data and provide AI-driven search and data augmentation?


There is a new generation of AI Native databases, which use deep learning models to find answers to natural language queries. We are talking about the ability to search through text, images, videos, DNA, or any unstructured data, all with a single query.


The rule of thumb: if there is an ML model, we can search through it.


Join me to learn about the foundation blocks (LLMs and vector embeddings, Vector Databases), how they all play together and most importantly - how you can build something yourself with open-source tech.


And, of course!!! There will be a live-coding demo, where I will take you through the experience of building an AI-based search – with Weaviate, an open-source Vector Database – and adding it to an app. Now the question... should this be done in Angular, React, Vue or just pure JS ;)


#MayTheDemoGodsBeWithUs


This talk has been presented at JSNation 2023, check out the latest edition of this JavaScript Conference.

FAQ

Machine learning (ML) is a field of artificial intelligence (AI) that focuses on building systems that learn from data, identify patterns, and make decisions with minimal human intervention. It involves training algorithms on data sets to create models that can perform tasks such as recognition, prediction, and decision-making.

Semantic search in machine and natural language processing involves understanding the intent and contextual meaning of a search query, rather than just matching keywords. It aims to improve the accuracy of search results by understanding the semantics, or meaning, of the words in the query.

Vector embeddings are mathematical representations of text or other data in a high-dimensional space. Machine learning models convert data into vectors of numbers that capture the essence of the data. These vectors are then used in various ML tasks, such as finding similarities between items or classifying data.

Machine learning has a wide range of applications including speech recognition, image recognition, medical diagnosis, predictive analytics, personalization in retail, and more. It is used to automate decision-making processes and create more personalized user experiences.

ChargPT appears to be a typographical error in the text, likely referring to 'ChatGPT', a variant of the GPT (Generative Pre-trained Transformer) models developed by OpenAI. These models use machine learning to generate human-like text based on the input they are given.

There are many tools available for implementing machine learning models, including TensorFlow, PyTorch, OpenAI's API, Hugging Face's Transformers, and others. These tools provide the infrastructure and pre-trained models that can be used to develop and deploy machine learning applications.

Search plays a crucial role in machine learning as it is often the starting point for gathering information and gaining insights. Improving search capabilities using ML can lead to more efficient data retrieval, better user experiences, and enhanced decision-making processes.

No, you do not need a PhD to understand machine learning. Many resources and tools are designed to make ML accessible to individuals without advanced degrees in the field.

Sebastian Witalec
Sebastian Witalec
31 min
01 Jun, 2023

Comments

Sign in or register to post your comment.
Video Summary and Transcription
The Talk discusses the use of machine learning in search engines, specifically focusing on semantic search and vector embeddings. It explores the integration of JavaScript and machine learning models, using Weaviate as an open-source vector database. The Talk demonstrates how to connect to Weaviate, query data, and perform machine learning queries. It also highlights the benefits of Weaviate, such as its superior developer experience and performance. Additionally, the Talk addresses customization options, data privacy concerns, and the varying effectiveness of different machine learning models.

1. Introduction to Machine Learning and Search

Short description:

I'm super excited to introduce you to this topic. I didn't have any idea about it a year ago. Let's get cracking. My journey begins with a conference talk on machine learning. At first, I was confused and thought I needed a PhD. But then I realized that machine learning tools have become accessible to everyone. I will focus on the topic of search using machine learning, as everything on the internet begins with search.

So, thank you for this very nice introduction. And I'm super excited to introduce you to this topic that, let's face it, a year ago I didn't have any idea about. So I'm going to bring you on my one year journey with me. So let's get cracking.

So you heard the introduction, my name is Sebastian Vitales, I build cool stuff, and I want to talk to you about it. So my journey begins a few years ago, a long long time ago in a galaxy far away. I went to a conference and there was a very interesting talk that I was very excited about that promised that you don't need a PhD to understand machine learning. And obviously I was super excited about it because A, I didn't have any clue about ML but I really wanted to get into it, like, hey, there was a promise, like, maybe I didn't need to study for five years to do something around ML. The thing is that 10 minutes into that talk I was so confused I didn't even know what my name was. And immediately my assumption was like, yep, you need a PhD, I'm never touching ML again in my life, you know. That's it. Over. But then I was like, OK, I'm not going to give up.

The thing that changed and is happening lately, and I'm sure you're all experiencing that, is that everybody's talking about ChargPT, AI, ML, like, all those things that you're hearing. Like, I currently live in Denmark. I don't speak any Danish but if I'm in a café there's, like, people speaking randomly Danish and catching ChargPT, something, something, like, you keep catching it. I went to get a haircut, right, like, and my hairdresser got confused. It was like, OK, I'm going to ask ChargPT what kind of haircut will go with you. Even ChargPT can help with it so that's OK. And the thing is, like, what changed? Why is everyone now talking again about machine learning? Where is this buzz coming from? Why is everyone excited, not just even people in tech, but even, like, regular people that, you know, don't, you know, don't use computers for professional stuff, right? And what changed is actually that those machine learning tools became accessible, right? Like, suddenly, they're, like, at your fingertips. Suddenly, you can go, you know, to OpenAI and, like, create a login, and then you could start writing prompts and ask questions to the AI and this is mind-blowing. And there are so many different applications, all sort of, like, image generation, all sort of things happening. But I only have 20 minutes for the talk. And the organizers, like, asked me already five times. So, I'm going to finish on time. So, I'm going to narrow down and only talk about, like, a very specific thing of the machine learning, especially I have 20 minutes and I want to do some live coding as well. So, let's stick to that.

So the topic of the presentation was search or using machine learning, search. And let's face it, everything that we do on the internet begins with search, right? You want to listen to music? You search. You want to watch a movie? You're going shopping? You want to find some information? You go to Wikipedia, you always search.

2. Challenges with Traditional Search

Short description:

Search works, but it could be better. Traditional search engines may not understand the meaning of a question, leading to irrelevant results. Semantic search, on the other hand, focuses on the meaning of the question and can provide more accurate answers. By using machine learning, we can enhance the power of semantic search.

And I mean, you probably think it's like, okay, what's the problem? We've been doing that for decades. Search works. Well, I beg to differ, right? Like, it kind of works, but it could be better. And let me give you an example. So, with traditional search, you may face some challenges. So, if you went and asked, like, a traditional search engine, like, why do airplanes fly? So, maybe you have a whole database of documents that explain it. You may get an answer like, why you should fly with expensive airplanes? And it's like, well, I mean, it's pretty good because it matches airplanes, it matches why do and fly and all this. Why is this guy complaining? It's like a perfect match. Well, I mean, in reality, we asked how planes fly, and we were told to fly with expensive airlines. Well, that solves it for me. Thank you. Well, the solution for me is kind of like looking for the question from the semantic point of view. What is the meaning of the question and what sort of answer can I find for you? And actually, if you put this question into Google Search, you'll get an answer like this. You go and we'll find the dynamics of fly from NASA. And then in there, the bit that helped us find the answer was like airplanes wings are shaped to make air move faster over the top of the wing, blah, blah, blah, blah, blah. Like we don't really have any of the keyword matches, but the meaning is there, right. And that's basically the power of semantic search. So by looking at those two examples and the kind of things that you can get between the two, I mean, the conclusion is pretty straightforward, right? Like we should be going and looking more at the semantic type of search and like using machine learning for it.

3. Machine Learning Models and Vector Embeddings

Short description:

In the machine learning world, we have machine learning models that can generate vector embeddings. These vector embeddings represent information and data in a multi-dimensional space. By passing a query through a machine learning model, we can obtain a vector that helps us find relevant articles. This process involves mathematical calculations and is accessible to people without a Ph.D. Additionally, some machine learning models can understand both text and images, allowing for searching through different types of data.

So how does this work? And a little heads up, you don't need to know it, but I think it's pretty cool. Like you'll find it interesting, and I have like a very nice explanation that you don't need a PhD to understand it. Maybe it's not four or five years old, like it's more of like a 10 years old explanation, so just bear with me.

So first of all, in the machine learning world, we have this thing called machine learning modes. There's this like black boxes that they are constructing in such way that if you pass in some information into it, it will spit out this thing which we call a vector embedding. A vector embedding is basically an array of numbers. It's just like a bunch of numbers, and in reality, it's like how computers look at data and you look like binary, and computers understand it, oh, this is a piece of information. This is basically what machine learning thinks. Like, if machine learning looks at the vector embedding, that's how it understands information and data.

So if we took our previous two examples that we had between the two results, and we put fly with expensive, as modern fluid paragraph into machine learning model, you will get a vector embedding. If you take another paragraph, and we put it again to machine learning model, we get another vector embedding. And what these things are, like you can imagine they are like coordinates in like a multi-dimensional space. Think of it like you see in this multi-dimensional space like dog and wolf that is right on the left-hand side. They're really close to each other, right? And a cat is not too far off, while the bananas and apples are in completely different place. And of course, our brains are trained to think in three dimensions. But those machine learning models can go into 300 dimensions, thousands, or 10,000 dimensions, so they can understand information in a very detailed way.

The cool thing is now that if I want to do a search, I could take our original query, pass it through the machine learning model, and I get a vector. With that vector, visually, to imagine it, we could say, like, hey, my information is somewhere here. So any objects that are close to that new vector position would indicate here are the articles where your answer might be. It's really cool, right? By looking there, we can then find, like, oh, yeah, the NASA article was just there, and then we can pass it back. It's all math. It's all just, like, grabbing and calculating those vectors, which is really, really cool. See? So I think even 10-year-olds and people that are not in ML can understand that you don't need a Ph.D. I think this is pretty straightforward.

So there's a bonus thing. Some of those machine learning models, and you probably saw there were some pictures, there was some text. Some of them can understand both text and images. They can also understand many other things. But in particular, there's one called clip, and if I pass in a paragraph that describes a maxi summer dress and a picture of the same dress separately, I would get two vectors which are very similar, which basically means that I could use it as well to search through images and text in however way we want to do. So I could have a database of pictures, and I could write in English kind of or any other language, so like, oh yeah, show me a picture with a house and a dog and a tornado in the background, right? And the vector would then help us find that picture later on.

4. JavaScript, Machine Learning Models, and Weaviate

Short description:

If we have it in the database, we could do images and text in one go. JavaScript plays a role in this. You need to pick up a machine learning model or vectorizer. Companies like Open AI, Cohere, Hug and Face, and Google have models available. Weaviate provides an open source vector database that handles everything for you. A demo showcases a database with 10 million objects, including paragraphs from Wikipedia in different languages.

Well, if we have it in the database. So yeah, we could do images and text in one go and plus many, many other things.

So this is a JavaScript event, and you're like probably beginning, like, OK, what does JavaScript got to do with it, right? So let me take you on that.

Now let's talk about the practice stuff. So if you have a JavaScript application, whether your search is done in the back end or the front end, usually you have some sort of JavaScript app, and you have data that you want to be able to search. Cool. That's the first ingredient we have.

Then what you need to do is pick up a machine learning model or a vectorizer, as we call them, and you don't necessarily have to create your own one. There's companies like Open AI. You probably know them because of ChargeAPT. There's Cohere, they have these multilingual models that understands over 100 different languages. We have Hug and Face. Hug and Face, they have 80,000 machine learning models. Recently, Google at Google I-O launched Palm 2, which also does this job. So you don't have to actually create those. You just basically find you need to find a model that works for your data, and you're golden. You're all good.

So having all this, to tie it all up together, and this is why I'm here. I'm wearing a t-shirt from Weaviate. We have an open source vector database. So it's basically a database that takes care of all of it for you, right? Like, you say, hey, I want to use this machine learning model. This is my data. Go. Right? And now I can write code in JavaScript or many other languages to write my queries or perform many other things.

So I want to do some cooking and show you a really cool demo. So, whoops. I was not going to do this. So the first demo I have is I have this database with 10 million objects in it. So these are paragraphs from Wikipedia. And to make it better, each million is a bundle of articles in different languages.

5. Connecting to WeaveIt and Querying Data

Short description:

With WeaveIt, I already have the data in my database. I connect to my instance in the cloud using the API key. Then, I create a client to write codes and queries. The builder pattern is used for querying. I query my articles to get text, URL, and title. I can also specify the number of objects to retrieve. Finally, I can perform a machine learning query using the 'with neartext' operator.

So I have some in English, I have some in French, in German, I have some in Hindi, in Arabic, I think there's Japanese as well. So we have like a pretty wide database. So what I can do with it is, I'm not going to be teaching you necessarily like, hey, this is React application, this is how you connect it to the UI, you know this stuff, right? But what you really need to know here is the following. So with WeaveIt, I already have the data in my database. So first thing that we need to do is connect to my instance. So this specific instance lives in the cloud. So we have this cloud services where in this case, I'm hosting this demo and in there I'm passing in the API key of my cloud instance. But additionally, I am using the Cohere API, like an inference service. So basically I'm telling like, hey, this is my API key. So if Cohere needs to charge me, they know how to find me, right? So I create this client and with this client, I can start writing codes and then writing my queries. So I'm not going to show you how to import the data in this case, I just want to focus on the querying part. So at the moment, we are using the builder pattern. So I'm grabbing a query from based on GraphQL. And then maybe we have with class. So my data is in collection called articles. Oh, I should actually probably show you this. So my collection articles is defined to use text of echo here. And then in this case, I want to use this multilingual model from 2022 December. And then also just so that you're aware, I have properties like text, title, URL, wiki ID, views, etc, etc. So now I'll be using those two to query through my data. So I want to query my articles and then we will query, what is it? Fields, and I want to get text, URL, and title. That's cool. And then maybe I want to just get first 10 objects. So if I just run this query, I could maybe search for how planes fly. At the moment, I'm not doing anything as such. I'm just saying, give me a bunch of random objects. OK, that's fine. But now I want to actually do a query. And this is the bit where I can do machine learning query with this one line. So I could say, with neartext, there's different operators as well.

6. Querying Concepts and Multiple Languages

Short description:

And I can pass in concepts. And the concept will be the query that I get from the UI. So this is the query I'm getting passed in. And now, with this one line, I can get articles that are related to flying. I can open this, and then this one is, what, is that Italian? Yep. So there's the Italian page on Wikipedia. And I found that article there. This is pretty cool. And of course, we can put many other queries.

And I can pass in concepts. And the concept will be the query that I get from the UI. So this is the query I'm getting passed in. And now, with this one line, I can get articles that are related to flying. I can open this, and then this one is, what, is that Italian? Yep. So there's the Italian page on Wikipedia. And I found that article there. This is pretty cool. And of course, we can put many other queries.

The other thing I can do is I can write it in other languages, right? Musiczne instrumenty. So this is in Polish, like musical instruments, right? And I'm searching across my database, which is super cool. And if I just want to filter and only show English response, so this is to show if you that we can also do traditional database kind of queries, so we can do scholar search. So I could do with, I need a dot. So with where. Actually I'm not going to type it manually, where language English. So in this case I'm just saying, okay, where language equals yen. So this time I can get English results only. Just pretty cool.

The other thing that I can do as like the next level, I have another collection where I needed to make some changes and I introduced a generative module model. So this one uses GPT from OpenAI. And I can actually do some extra bits and then do, like provide prompts to, to the database to, to perform some extra operations. So let's try that. So this time the collection is called Wikipedia, but it's basically the same data. I just didn't have the time to rebuild the other one. So we're going to skip this one for now and then maybe I can do with generate. And then in this case, we need, what do we need? Show prompt. And then I can say summarize the following in a tweet. And then I'm going to pass text, which is this property. And because I'm using OpenAI this time, I am going to add an API key. So I have it in my local configuration.

7. Running Queries and Image Search

Short description:

Let's not run on 10 because that's a bit time consuming. We run the query, get the results, pass it on to OpenAI to vectorize. I summarize the following as a haiku poem. We have haiku poems based on the article with Galswin. The next demo is image search. I have a folder with famous actors and celebrities. I use a clip model to find a celebrity that looks like Brad Pitt. I return two objects, such as David Tennant and Stan Tucci. I also did the same thing for the conference speakers, Jeff Cross and Mike Harrington.

Let's not run on 10 because that's a bit time consuming. What will happen is we run the query, we get the results, and then we'll pass it on to OpenAI to vectorize. And then also I need to do a little bit of parsing. So generate single result. Nope, not this one. Demo. Return. Generate. It's just like a simple map. I'm just after this field and I will replace the text. So if I rerun this and run the search, that's going to take, usually takes 10, 15 seconds, but yeah.

It takes each of the results and then creates a summarizer. What's really cool, everything's super cool when I'm talking right now, I summarize the following as a haiku poem. I save this and then I can rerun this and this is mind blowing. So it takes a while. And we have haiku poems based on the article with Galswin. All right. The next demo I have for you is, you could also do image search. So I have this table, collection, whatever you want to call it where, well, this is going to be tiny but you should be able to see now. So I have like a folder with like a bunch of actors, right? Famous actors, famous celebrities and this time I'm using a clip model. So I could, for example, do this where I say find a celebrity that looks like Brad Pitt. And then, boom! We have Brad Pitt but I'm always going to return two objects so not all of them would be perfect. Which actor do I look like from the like a hundred that I have? I look like David Tennant and Stan Tucci. They're really going after my Baldy here. This is offensive. But hey, another one, like I have a picture of a meerkat. So well actually Ryan Gosling does look a bit like a meerkat so that works pretty nicely. And I did the same thing for the conference speakers here. There's a bit of a running joke when it comes to Jeff Cross and Mike Harrington. They usually get mixed up.

8. Building an AI-based Search Application

Short description:

At one point they even swapped their profile pictures. This is really, really powerful. You can load them into Weave8 and then you run them through a machine learning model and then query them. As a bonus, our website uses vector search. So to summarize, anyone can bring AI-based search to their web application.

At one point they even swapped their profile pictures. If I show you again the picture of Jeff, you know, there's some resemblance between them too. And if you want to search for Mishko, there was a different picture of Mishko. We're still able to find it. And this is really, really powerful.

And this application I built in Angular. So just to show you the little bit on how this is done. So we have this image service. So basically the step is this time I'm using the multi, like the clip model for the vectorization or to understand our images. And I'm saying, like, hey, we'll have name URL and image. And then after, then I go and import the data. This is not super important. The image search is just this. Basically, when you give me a file to search on, I convert it to base 64. And then the query is just like the one you saw before, with the difference being that now I'm doing with near image. So I'm passing this new image as a base 64 representation, and boom, I can do the search. And you see, like, none of this code makes you, like, requires you to understand how machine learning works underneath. You just have a job to do.

And so if in your websites or in your backend systems, maybe, I don't know, you have 500 million PDFs that your company has been storing for the last 50 years, and then you could never search through them. Like, you can load them into Weave8 and then you run them through a machine learning model and then query them. Like, something that was inaccessible before, now you can run it. As a bonus, our website, we've added this recently. So if I do Command K, we actually use vector search in our website itself. How to delete an object, boom, here is an article in a documentation. So these are all really, really cool things. So to summarize, anyone can bring AI-based search to their web application. Basically, you need your JS application, WebSum data, pick a machine learning model, and a vector database. And with that, and then, what, like there was like 20 or 30 lines of code, you're golden, and then you can use machine learning search for your use. So yeah, that's me. This is our website.

9. Benefits and Differences with Pinecone

Short description:

WeaveIt is an open source database that can be deployed in the cloud or run with Docker and Kubernetes. The developer experience is superior, allowing for running queries with just five lines of code.

This is our Twitter. So the first question we have, and so I'm not really familiar with this space, but they're asking what are the benefits or maybe the differences between a product called Pinecone? Oh, yeah. The Pinecone would be like our competitor. I think one of the biggest differences, like first of all, WeaveIt is an open source database and we offer you both that you can deploy into the cloud, but also the open source option, you can run them with Docker and with Kubernetes. But the other thing also is that our developer experience, I think is superior. So like the fact that like I was able to run a query, like five lines of code, that's not possible with any other vector database that's out there.

10. Comparison Algorithm and Performance

Short description:

The comparison algorithm used in Weaviate is based on a special indexing mechanism called HNSW. This mechanism allows for fast searching of complex and long vectors. Weaviate has been tested on nearly a billion objects, with queries running under 100 milliseconds even with 960 million objects. It's truly impressive.

Right, that's a good point. And this next one is asking about the comparison algorithm, but I guess that has to do with how it's actually dealing with those vector embeddings. So they're asking how does that comparison algorithm work and how performant is it? Yeah, how it works is actually, I could give like a one hour talk about it, by the way, but basically the idea is that because the vectors are like very complex, they're long and you can't just sort them by one thing. So we just have a special indexing mechanism that allows you to search very fast. One of the first ones implemented is called HNSW. It's very technical, but the idea is we just have a specific index mechanism for it. How performant it is? We've tested Weaviate on nearly a billion objects, which topped out 960 million because we run out of data. Even with this 960 million we're running queries under 100 milliseconds, which is insane because of the vast amount of data. Yeah, it's crazy.

QnA

Customizing Distance Metric and AI Search

Short description:

This next one asks if you can customize the distance metric. Out of the box, Weaviate supports multiple distance metrics, and it's easy to add new ones. Another question is about using Postgres for vector embedding. While it's possible, Postgres may struggle with large datasets, unlike Weaviate. Training the AI in Weaviate doesn't take much time, as the machine learning model used is already trained. Importing data and converting it into vector embeddings is the main investment of time. When it comes to indexing PDF data, Weaviate is working on it and has published a blog post explaining the process. AI search can be useful for deterministic search, but it depends on the specific case.

This next one asks can you customize the distance metric? So like cosine similarity or Manhattan, is that something you would do? Yup, so basically out of the box we support five or six different distance metrics, but also Weaviate A is very modular and open source. It's very easy to add new ones.

This next question is asking about Postgres. So they mentioned Postgres has vector embedding, so I guess why would you use something like this if something like Postgres already support something similar? That's the thing. I could technically take a Formula One car on the street and then try to drive around. And it's all about is the tool built for the job that I'm trying to do? And when it comes to vector embedding search, there's no comparison. I think you can do it in Postgres, but the moment you hit like half a million objects, you will really struggle. In our case, I'm boasting and we go in a billion, and we're running out of data. And I think that's the big thing.

This next one is how much time is needed to train the AI? For example, on the Wiki. And I guess to test my own understanding, is it basically the time it takes to generate the vector embeddings? And at that point, you're ready to go? Yeah, so actually, it's not so much as a training, because the machine learning model that we use with Cohere there is already trained. So, the only time that you have to invest is as you import the data. All that data has to be converted into those vector embeddings. I think I imported those 10 million objects, it took me a couple of hours to do it from my laptop on a crappy network in a local café. That's crazy. So, you know, it's doable, it's not too bad. But there's an investment of time upfront required.

This next one asks, so what if you want to index like PDF data? Is it a similar process? It's slightly different, but actually we are working on it. We published a blog post about it last week, Tuesday. So, go to our website and check out our blogs. But yeah, we're dealing with unstructured. I think the trick is basically have something that can read through those PDFs, and maybe construct paragraphs, and then pass them into Weave 8. Because one of the things that's worth understanding, if you pass in, like, let's say a whole book into a machinery model, it will speed up one vector. Like and then if you run a query, it's just like, yeah, that's an average of the whole Bible, the whole thing. So, what you really want is something that can find specific parts. So, that chunking is sometimes usually the process that takes a little bit of extra effort. But, yeah, that blog post actually we explained pretty well. So, it's more about like, can I read through it, and then chunk up my content in a nice way? Right.

This next one is, is AI search useful for search that is completely deterministic or is it overkill? I think it depends on the case-by-case.

AI Search and Training Models

Short description:

AI search can be useful in certain cases, but it depends on the specific situation. Weaviate offers both AI search and hybrid search, allowing users to choose between traditional methods and AI search. The developer experience is seamless, with a simple method call to decide whether to use AI or not. The amount of data needed for training models and creating embeddings varies, but generally, more data leads to better results. However, there is no definitive answer to how much data is enough. Each model has its own unique vectors, and they are not interchangeable.

This next one is, is AI search useful for search that is completely deterministic or is it overkill? I think it depends on the case-by-case. And actually to that point, that's why in Weaviate we offer like we have AI search but we also have hybrid search where you can use both like the traditional methods and AI search in one go. And we also support this algorithm called BM-25, so that's keyword based search, so you could also do that. So based on what you need, you could do either.

Yes, there's cases where it just doesn't make sense, right, where you have like you know there's this exact thing you're looking for, you don't need AI for it. So actually the search that I showed on the website when I was searching for how to delete an object, we use hybrid there because in some cases the ML model might not know about some function names that we use. And I think just looking at the developer experience, it's almost like with just a method call you decide, do I want to use AI or not? So that's pretty cool.

By the way, I built this demo, the image search demo in like two hours like three days ago. I was just like in the evening, I was like, oh yeah, I want to add another demo and it just took me two hours, and I built it from scratch including installing Angular on my laptop and all this. Nice. Would it be possible to train a model on, for example, hundreds of product manuals like phones or TVs and get accurate results, or is a larger amount needed? So I guess for anything you're training or creating the embeddings for, how much data do you need? Yeah, that would be a good question to somebody in my team. I mean like basically what I'm being told is like the more data you have, the better quality results you can get. So I think like what you see with ChargeGPT, they had enough data that they had to spend six months training it, so we're talking about crazy amounts. So it's all about like the more data you have, the better it is. But what's enough? That's kind of hard depending on how good the results you want. I don't have a good answer for it, I'm sorry. Okay, no worries. I don't think there is one. Yeah. I guess it just depends. You have to try it out and see. Yeah. The next one is, if you get vector data out of something like text data 002, is it going to output the same vectors as some other model? Are they interchangeable or? No. So each model has its own thing. So the way I see this, I use this as an example. Imagine you have somebody that works at the shopping mall and knows where all the stores are, all the shops are, and then you go, like, hey, you know, like, coffee. And then he goes, like, this place, this place, right? And now you take him to Amsterdam and put him in a different one, and you're like, coffee shop is like this way, this way. He'll be sending you wrong directions, right? So that's like, what are they trained on, right, that matters. Yeah.

Embeddings, Adding Vectors, and Data Privacy

Short description:

Have you tried any other embeddings like GloVe or Word2Vec? How much better is it than Transformers? Is it always worth it? It's easy to test it out in Wavi8. Does it take a lot of time to add vectors for big data? The Wikipedia demo runs on an instance in the cloud, while the image search runs on a laptop. Vector databases are RAM hungry, and ChatGPT uses input queries to train its future requests. If you're concerned about data privacy, avoid putting sensitive information on the internet.

Yeah. Makes sense. Have you tried any other embeddings as, like, GloVe, Word2Vec? Don't know what that is but maybe you do. And how much better is it than Transformers? Is it always worth it? Imagine the cost of learning. Yeah. That's not a good question for me, honestly. Like, yeah, this will be for, like, the data scientists on my team.

Okay. I can definitely say for sure that it's easy to test it out in Wavi8, so I'll just encourage swapping them out and trying out. Yeah. Yeah, and I think we sort of covered on this one earlier but does it take a lot of time to add vectors for big data? But I guess you mentioned you were able to train on your laptop, so... Yeah. I mean, so the Wikipedia demo that doesn't run on my laptop, that actually we have an instance in the cloud. Okay. The image search that I did, that was purely on my laptop.

Yeah. I think the biggest thing that you need when you think of vector database that we don't say, it's, like, it's very RAM hungry. Mm. And the reason that index wide runs super fast is because we store it in RAM. I see. Yeah. Okay. And I guess the last one to kind of end on, the morals of it all, so should we be concerned about sharing our app data with an AI tool like ChatGPT? Yeah, I mean, so there is the thing of, like... So one thing that I know if you send queries to ChatGPT, they do use the input to train their future, basically, requests. And especially if you, you know, using ChatGPT for free, it's kind of like, it's an exchange. You give them training data in exchange for what you do. And I think recently Apple forbade their employees to use ChatGPT for any content. So there might be something to it. But generally, there is no malicious intent as such. We're friends, we're actually people at OpenAI. But if you're really concerned and say, hey, only I and another person should ever see this piece of information, like, yeah, just don't put it on the internet, right? Awesome. So that's all we have for you. Thank you so much, Sebastian. Thank you.

Check out more articles and videos

We constantly think of articles and videos that might spark Git people interest / skill us up or help building a stellar career

Building a Voice-Enabled AI Assistant With Javascript
JSNation 2023JSNation 2023
21 min
Building a Voice-Enabled AI Assistant With Javascript
Top Content
This Talk discusses building a voice-activated AI assistant using web APIs and JavaScript. It covers using the Web Speech API for speech recognition and the speech synthesis API for text to speech. The speaker demonstrates how to communicate with the Open AI API and handle the response. The Talk also explores enabling speech recognition and addressing the user. The speaker concludes by mentioning the possibility of creating a product out of the project and using Tauri for native desktop-like experiences.
The Ai-Assisted Developer Workflow: Build Faster and Smarter Today
JSNation US 2024JSNation US 2024
31 min
The Ai-Assisted Developer Workflow: Build Faster and Smarter Today
Top Content
AI is transforming software engineering by using agents to help with coding. Agents can autonomously complete tasks and make decisions based on data. Collaborative AI and automation are opening new possibilities in code generation. Bolt is a powerful tool for troubleshooting, bug fixing, and authentication. Code generation tools like Copilot and Cursor provide support for selecting models and codebase awareness. Cline is a useful extension for website inspection and testing. Guidelines for coding with agents include defining requirements, choosing the right model, and frequent testing. Clear and concise instructions are crucial in AI-generated code. Experienced engineers are still necessary in understanding architecture and problem-solving. Energy consumption insights and sustainability are discussed in the Talk.
AI and Web Development: Hype or Reality
JSNation 2023JSNation 2023
24 min
AI and Web Development: Hype or Reality
Top Content
This talk explores the use of AI in web development, including tools like GitHub Copilot and Fig for CLI commands. AI can generate boilerplate code, provide context-aware solutions, and generate dummy data. It can also assist with CSS selectors and regexes, and be integrated into applications. AI is used to enhance the podcast experience by transcribing episodes and providing JSON data. The talk also discusses formatting AI output, crafting requests, and analyzing embeddings for similarity.
The Rise of the AI Engineer
React Summit US 2023React Summit US 2023
30 min
The Rise of the AI Engineer
Top Content
Watch video: The Rise of the AI Engineer
The rise of AI engineers is driven by the demand for AI and the emergence of ML research and engineering organizations. Start-ups are leveraging AI through APIs, resulting in a time-to-market advantage. The future of AI engineering holds promising results, with a focus on AI UX and the role of AI agents. Equity in AI and the central problems of AI engineering require collective efforts to address. The day-to-day life of an AI engineer involves working on products or infrastructure and dealing with specialties and tools specific to the field.
Web Apps of the Future With Web AI
JSNation 2024JSNation 2024
32 min
Web Apps of the Future With Web AI
Web AI in JavaScript allows for running machine learning models client-side in a web browser, offering advantages such as privacy, offline capabilities, low latency, and cost savings. Various AI models can be used for tasks like background blur, text toxicity detection, 3D data extraction, face mesh recognition, hand tracking, pose detection, and body segmentation. JavaScript libraries like MediaPipe LLM inference API and Visual Blocks facilitate the use of AI models. Web AI is in its early stages but has the potential to revolutionize web experiences and improve accessibility.
Code coverage with AI
TestJS Summit 2023TestJS Summit 2023
8 min
Code coverage with AI
Premium
Codium is a generative AI assistant for software development that offers code explanation, test generation, and collaboration features. It can generate tests for a GraphQL API in VS Code, improve code coverage, and even document tests. Codium allows analyzing specific code lines, generating tests based on existing ones, and answering code-related questions. It can also provide suggestions for code improvement, help with code refactoring, and assist with writing commit messages.

Workshops on related topic

AI on Demand: Serverless AI
DevOps.js Conf 2024DevOps.js Conf 2024
163 min
AI on Demand: Serverless AI
Top Content
Featured WorkshopFree
Nathan Disidore
Nathan Disidore
In this workshop, we discuss the merits of serverless architecture and how it can be applied to the AI space. We'll explore options around building serverless RAG applications for a more lambda-esque approach to AI. Next, we'll get hands on and build a sample CRUD app that allows you to store information and query it using an LLM with Workers AI, Vectorize, D1, and Cloudflare Workers.
AI for React Developers
React Advanced 2024React Advanced 2024
142 min
AI for React Developers
Top Content
Featured Workshop
Eve Porcello
Eve Porcello
Knowledge of AI tooling is critical for future-proofing the careers of React developers, and the Vercel suite of AI tools is an approachable on-ramp. In this course, we’ll take a closer look at the Vercel AI SDK and how this can help React developers build streaming interfaces with JavaScript and Next.js. We’ll also incorporate additional 3rd party APIs to build and deploy a music visualization app.
Topics:- Creating a React Project with Next.js- Choosing a LLM- Customizing Streaming Interfaces- Building Routes- Creating and Generating Components - Using Hooks (useChat, useCompletion, useActions, etc)
Vibe coding with Cline
JSNation 2025JSNation 2025
64 min
Vibe coding with Cline
Featured Workshop
Nik Pash
Nik Pash
The way we write code is fundamentally changing. Instead of getting stuck in nested loops and implementation details, imagine focusing purely on architecture and creative problem-solving while your AI pair programmer handles the execution. In this hands-on workshop, I'll show you how to leverage Cline (an autonomous coding agent that recently hit 1M VS Code downloads) to dramatically accelerate your development workflow through a practice we call "vibe coding" - where humans focus on high-level thinking and AI handles the implementation.You'll discover:The fundamental principles of "vibe coding" and how it differs from traditional developmentHow to architect solutions at a high level and have AI implement them accuratelyLive demo: Building a production-grade caching system in Go that saved us $500/weekTechniques for using AI to understand complex codebases in minutes instead of hoursBest practices for prompting AI agents to get exactly the code you wantCommon pitfalls to avoid when working with AI coding assistantsStrategies for using AI to accelerate learning and reduce dependency on senior engineersHow to effectively combine human creativity with AI implementation capabilitiesWhether you're a junior developer looking to accelerate your learning or a senior engineer wanting to optimize your workflow, you'll leave this workshop with practical experience in AI-assisted development that you can immediately apply to your projects. Through live coding demos and hands-on exercises, you'll learn how to leverage Cline to write better code faster while focusing on what matters - solving real problems.
Building Full Stack Apps With Cursor
JSNation 2025JSNation 2025
46 min
Building Full Stack Apps With Cursor
Featured Workshop
Mike Mikula
Mike Mikula
In this workshop I’ll cover a repeatable process on how to spin up full stack apps in Cursor.  Expect to understand techniques such as using GPT to create product requirements, database schemas, roadmaps and using those in notes to generate checklists to guide app development.  We will dive further in on how to fix hallucinations/ errors that occur, useful prompts to make your app look and feel modern, approaches to get every layer wired up and more!  By the end expect to be able to run your own AI generated full stack app on your machine!
Please, find the FAQ here
Free webinar: Building Full Stack Apps With Cursor
Productivity Conf for Devs and Tech LeadersProductivity Conf for Devs and Tech Leaders
71 min
Free webinar: Building Full Stack Apps With Cursor
Top Content
WorkshopFree
Mike Mikula
Mike Mikula
In this webinar I’ll cover a repeatable process on how to spin up full stack apps in Cursor.  Expect to understand techniques such as using GPT to create product requirements, database schemas, roadmaps and using those in notes to generate checklists to guide app development.  We will dive further in on how to fix hallucinations/ errors that occur, useful prompts to make your app look and feel modern, approaches to get every layer wired up and more!  By the end expect to be able to run your own ai generated full stack app on your machine!
Working With OpenAI and Prompt Engineering for React Developers
React Advanced 2023React Advanced 2023
98 min
Working With OpenAI and Prompt Engineering for React Developers
Top Content
Workshop
Richard Moss
Richard Moss
In this workshop we'll take a tour of applied AI from the perspective of front end developers, zooming in on the emerging best practices when it comes to working with LLMs to build great products. This workshop is based on learnings from working with the OpenAI API from its debut last November to build out a working MVP which became PowerModeAI (A customer facing ideation and slide creation tool).
In the workshop they'll be a mix of presentation and hands on exercises to cover topics including:
- GPT fundamentals- Pitfalls of LLMs- Prompt engineering best practices and techniques- Using the playground effectively- Installing and configuring the OpenAI SDK- Approaches to working with the API and prompt management- Implementing the API to build an AI powered customer facing application- Fine tuning and embeddings- Emerging best practice on LLMOps