Imagine having an AI voice assistant like Iron Man's J.A.R.V.I.S., capable of understanding and responding to your needs in real-time. In this guide, you will learn how to build your very own real-time AI voice assistant using ElevenLabs, store each conversation in a Postgres database, and index them for faster retrieval.
Deal alert
Take advantage of the AI Engineer Starter Pack by ElevenLabs to get discounts for the tools used in this guide.
Let’s get started by creating a new Next.js project with the following command:
When prompted, choose:
Yes when prompted to use TypeScript.
No when prompted to use ESLint.
Yes when prompted to use Tailwind CSS.
No when prompted to use src/ directory.
Yes when prompted to use App Router.
No when prompted to use Turbopack for next dev.
No when prompted to customize the default import alias (@/*).
Once that is done, move into the project directory and install the necessary dependencies with the following command:
The libraries installed include:
framer-motion: A library for animations in React.
react-feather: A collection of open-source icons for React.
motion: A library to create animations in React applications.
sonner: A notification library for React to display toast notifications.
@11labs/react: A React library to interact with ElevenLabs API.
@neondatabase/serverless: A library to connect and interact with Neon’s serverless Postgres database.
The development-specific libraries include:
tsx: To execute and rebuild TypeScript efficiently.
Provision a Serverless Postgres
To set up a serverless Postgres, go to the Neon console and create a new project. Once your project is created, you will receive a connection string that you can use to connect to your Neon database. The connection string will look like this:
Replace <user>, <password>, <endpoint_hostname>, <port>, and <dbname> with your specific details.
Use this connection string as an environment variable designated as DATABASE_URL in the .env file.
Create an AI Agent with ElevenLabs
To create a customizable agent, go to ElevenLabs' AI Agents and then click on Create an AI agent button.
Next, give it a personalized name and select the kind of Agent you would want. For demonstration purposes, let's start with a Blank template.
Next, copy the Agent ID displayed just below the customized name of your agent (here, Custom). You will use this Agent ID as the AGENT_ID environment variable in your application.
Next, go to Advanced > Client Events in your Agent settings, and add two events named agent_response and user_transcript.
Finally, go to API Keys, create an API key and use the value obtained as XI_API_KEY enviroment variable in your application.
Database Schema Setup
Create a file named schema.tsx at the root of your project directory with the following code:
The code above defines an asynchronous function createMessagesTable that connects to a Neon serverless Postgres database using a connection string stored in the DATABASE_URL environment variable, creates a messages table if it doesn't already exist, and sets up an index on the session_id and created_at columns for faster retrievals.
To run the migrations, execute the following command:
If it runs succesfully, you should see Setup schema succesfully. in the terminal.
Build Reusable React Components and Hooks
1. Typing Effect Animation
To enhance the user experience by simulating real-time interactions, implement a typing effect in the UI to render AI responses incrementally. Create a file named useTypingEffect.ts in the components directory with the following code:
The provided code exports a custom React hook called useTypingEffect. This hook simulates a typing effect for a specified text over a given duration, enhancing the user interface by rendering text incrementally.
2. Conversation Message
To render each message in the conversation history, you need to dynamically indicate whether the message is from the User or the AI. Create a file named Message.tsx in the components directory with the following code:
The code above exports a React component that renders a message. It conditionally displays a Cpu icon for messages from the AI and a User icon for messages from the user, along with the message content.
3. Various States During AI Interaction
Create a file named TextAnimation.tsx in the components directory with the following code:
The code above exports a React component that creates an interactive UI for the AI voice assistant. It utilizes the useTypingEffect hook to simulate a typing effect for the AI's responses and displays different states of interaction, such as "idle," "listening," and "speaking." The component also includes a clickable circle that toggles between starting and stopping the listening state, providing visual feedback through animations.
Generate a Signed URL for private conversations with ElevenLabs
To create a secure access between user and AI (powered by ElevenLabs), create a new file named route.ts in the app/api/i directory with the following code:
The code above defines an API route that generates a signed URL using ElevenLabs API. You will want to use signed URL instead of connecting to a fixed point server so as to allow connection to your personalized, private agents created in ElevenLabs.
Sync Conversations to a Postgres database
Create a file named route.ts in the app/api/c directory with the following code:
The code above defines two endpoint handlers on /api/c:
A POST endpoint that allows you to insert a new message into the messages table. It expects a JSON payload containing the id of the session and the item to be inserted. If the session ID or item is missing, it returns a 400 status code.
A GET endpoint that retrieves all messages associated with a specific session ID. It extracts the session ID from the request URL and queries the messages table, returning the results as a JSON response. If the session ID is not provided, it returns an empty array.
Create the UI for Starting Conversations and Synchronizing Chat History
Create a file named page.tsx in the app/c/[slug] directory with the following code:
The code above does the following:
Defines a loadConversation function which calls the /api/c route to fetch the conversation history based on the particular slug (i.e. the conversation ID).
Uses the useConversation hook by ElevenLabs to display the toast when the instance is connected, and to sync the real-time message to Postgres using the onMessage callback.
Defines a connectConversation function that instantiates a private conversation with the agent after obtaining a signed URL using the /api/i route.
Defines a disconnectConversation function that disconnects the ongoing conversation with the agent.
Creates a useEffect handler which on unmount, ends the ongoing conversation with the agent.
Next, import the TextAnimation component which displays different state of the conversation, whether AI is listening or speaking (and what if so).
Finally, add a Show Transcript button that displays the conversation history stored in Neon to the user.
Now, let's move on to deploying the application to Vercel.
Deploy to Vercel
The repository is now ready to deploy to Vercel. Use the following steps to deploy:
Start by creating a GitHub repository containing your app's code.
Then, navigate to the Vercel Dashboard and create a New Project.
Link the new project to the GitHub repository you've just created.
In Settings, update the Environment Variables to match those in your local .env file.
In this guide, you learned how to build a real-time AI voice assistant using ElevenLabs and Next.js, integrating it with a Postgres database to store and retrieve conversation histories. You explored the process of setting up a serverless database, creating a customizable AI agent, and implementing a user-friendly interface with animations and message handling. By the end, you gained hands-on experience connecting various technologies to create a fully functional AI voice assistant application.
Need help?
Join our Discord Server to ask questions or see what others are doing with Neon. Users on paid plans can open a support ticket from the console. For more details, see Getting Support.