Create an AI Chat App in Minutes
Note: This is an activity that we do with second year students. They are quite proficient in JavaScript, but if you know just a little bit about NodeJS you should be able to set this up and get it working.
We'll be using a Google Gemini AI model to power this chat app, so the first step is to get an API key from Google. Luckily, they've made this very easy to do. All you have to do is:
- Go to Google's AI Studio
- If you've never visited this site, you'll be prompted to accept the terms of service.
- Click the Get API Key button in the upper left corner
- Then click the Create API Key button If you've never created an app on the Google Cloud Platform, this will create one along with an API key that you can use for this activity. Make sure to store the key in a safe place.
Now create a folder named node-ai-chatbot, then open the folder in VSCode.
Initialize the project by entering this command:
npm init -y
We'll use ES6 modules for this project, so add this setting to your package.json file:
"type": "module"
Now install the dependencies:
npm install express dotenv @langchain/google-genai
Create an .env file in the project folder and put this into it (but use your YOUR api key):
GEMINI_API_KEY="YOUR-API-KEY-GOES-HERE"
Create a file named app.js in the project folder, and put this into it:
import dotenv from 'dotenv';
import { ChatGoogleGenerativeAI } from "@langchain/google-genai";
import express from 'express';
const app = express();
dotenv.config()
const model = new ChatGoogleGenerativeAI({
model: "gemini-1.5-flash",
maxOutputTokens: 2048,
apiKey: process.env.GEMINI_API_KEY,
temperature: 0.7,
system_message: "Do not use Markdown." // It's still using markdown!!!!
});
app.use(express.static('public'));
app.use(express.json());
app.post("/send-context", async (req, res, next) => {
const {history} = req.body;
const result = await model.invoke(history);
res.json(result.content);
});
const port = 8888
const server = app.listen(port, () => {
console.log("Waiting for requests on port %s", port);
});
Create a public folder in the project folder, then create a file called index.html inside the public folder. Put this code inside of it:
<!DOCTYPE html>
<html lang="en">
<head>
<meta charset="UTF-8">
<meta name="viewport" content="width=device-width, initial-scale=1.0">
<title>ChatBot</title>
<script src="https://cdnjs.cloudflare.com/ajax/libs/axios/1.7.7/axios.min.js" integrity="sha512-DdX/YwF5e41Ok+AI81HI8f5/5UsoxCVT9GKYZRIzpLxb8Twz4ZwPPX+jQMwMhNQ9b5+zDEefc+dcvQoPWGNZ3g==" crossorigin="anonymous" referrerpolicy="no-referrer"></script>
</head>
<body>
<h3>Enter your prompt:</h3>
<form id="chat-form">
<input type="text" name="message" style="width: 350px;">
<input type="submit" value="Send">
</form>
<ul id="conversation-list"></ul>
<script>
// Configure the Axios library to send the conversation to the server
const ax = axios.create({
baseURL: 'http://localhost:8888',
headers: {'Content-Type': 'application/json'}
});
// The function that uses Axios to send the conversation to the server
const sendConversation = async (conversation) =>{
const requestBody = {
history: conversation
}
const resp = await ax.post("/send-context", requestBody);
return resp.data;
}
// The function that will display the conversation on the page
// The msgArray param might look like this: ["human", "Hello!"]
// OR like this: ["assistant", "Hi there!"]
const displayMessage = (msgArray) => {
const li = document.createElement("li");
const h3 = document.createElement("h3");
const p = document.createElement("p");
h3.textContent = msgArray[0]; // "human" or "assistant"
p.textContent = msgArray[1]; // The message from the human or assistant(AI model)
li.append(h3, p);
conversationList.append(li);
}
// Get the HTML elements we need from the page
const form = document.querySelector("#chat-form");
// The textbox for the human message:
const txtMessage = form.querySelector("[name=message]");
// This list (OL element) that will display the conversation:
const conversationList = document.querySelector("#conversation-list");
// Set the system prompt before starting the conversation:
let conversation = [
["system", "You are an expert at answering general questions in a concise and easy-to-understand manner."]
];
// handle form submits, for when the user/human types a message
form.addEventListener("submit", async (evt) => {
evt.preventDefault();
// Get the user/human input and put it into an array
const message = txtMessage.value;
const humanMessage = ["human", message];
// Display the human message in the conversation list
displayMessage(humanMessage);
txtMessage.value = "";
// add the human message to the conversation array
conversation.push(humanMessage);
// send the conversation to the server and wait for the AI response
const aiResponse = await sendConversation(conversation);
// Add the ai response to the conversation array
const aiMessage = ["assistant", aiResponse];
conversation.push(aiMessage);
// Display the AI message in the conversation list
displayMessage(aiMessage);
txtMessage.focus();
});
</script>
</body>
</html>
Start the app by the app by running this command (from the node-ai-chatbot folder):
node app.js
Then navigate to localhost:8888 in your browser, and you should be able to start a conversation with the Gemini LLM!
How It Works - Understanding how the context data is stored
To understand how the code is working, you need to know how you should structure the data for the conversation. We are using the LangChain library to interact with Gemini, so it expects you to store the conversation (the context) in a multidimensional array. It could start out looking like this:
[
["system", "Your responses should be clear and concise"]
]
The first element in this array is an array that has two strings in it. The first of which is "system", and the second of which tells the AI model how it should respond to your prompts throughout the conversation. The 'system' prompt is not required, but it's common to provide some guidance to the AI model in how it should respond to your prompts. For example, you may want to reply with a humorous tone, or you may want to give it some information about like "explain things in a way that a 5th grader can understand them".
After the 'system' prompt, you can begin the conversation telling the model what you want. Your input will be marked as 'human', and the responses from the AI model will be marked as 'assistant'.
So, to get the conversation started, you could another array that includes your initial prompt, maybe something like this:
[
["system", "Your responses should be clear and concise"],
["human", "Help me name my new puppy"]
]
Now we are ready to send the data to the AI model to get the conversation going, then we'll wait for it to add it's response.
Axios is used to send the initial data from the web page to the node server, and the node server uses the LangChain library to send it to the Gemini model.
When the model responds, the node server will add the response to the context/conversation, like so:
[
["system", "Your responses should be clear and concise"],
["human", "Help me name my new puppy"],
["assistant", "How about Old Yeller?"]
]
This data is then sent back to the client (the web page), which will display the entire conversation/context in the web page. You might then enter a prompt, which will update the context data to something like this:
[
["system", "Your responses should be clear and concise"],
["human", "Help me name my new puppy"],
["assistant", "How about Old Yeller?"],
["human", "Can you suggest a different name?"]
]
And the process would continue so that each time you enter a prompt, an array with the first element of "human" will be added to the data. And when the AI model responds, and array with the first element of "assistant" will be added.
This is how the model maintains the context of the conversation, it can't truly remember the conversation on its own. So you have to keep sending the entire context so that it knows what you've already discussed.