Skip to content

An example of streaming ChatGPT via the OpenAI v4.0 node SDK.

License

Notifications You must be signed in to change notification settings

sshader/streaming-chat-gpt

 
 

Repository files navigation

Streaming Chat Completions With HTTP Actions

An example of streaming ChatGPT via the OpenAI v4.0 node SDK built off of this repo with this accompanying Stack post.

However this app uses an HTTP endpoint with response streaming to send the ChatGPT response to the message author without having to update the database on every streamed part.

The result is that the author of a message sees frequent updates to their message, while any other viewers receive batched updates, and the app overall requires less database bandwidth.

Diagram of browsers talking to Convex, which talks to OpenAI

Demo of message streaming from OpenAI with two different users

Overview:

This app makes up a name for your user using Faker.js and allows you to chat with other users: open up multiple tabs to try it out!

If your message includes "@gpt" it will kick off a request to OpenAI's chat completion API and stream the response, updating the message as data comes back from OpenAI.

  • The frontend logic is all in App.tsx.
  • The server logic that stores and updates messages in the database is in messages.ts.
  • The HTTP endpoint that makes the streaming request to OpenAI is in http.ts.
  • The initial messages that are scheduled to be sent are in init.ts.

Running the App

npm install
npm run dev

This will configure a Convex project if you don't already have one. It requires an OpenAI API key. Set the environment variable: OPEN_API_KEY (should start with sk-) in your Convex backend via the dashboard once your backend has been configured. You can also get there via:

npx convex dashboard

Once npm run dev successfully syncs, if the database is empty, it will schedule some messages to be sent so you can see it working in action.

It will then start two processes in one terminal: vite for the frontend, and npx convex dev for syncing changes to Convex server functions.

Check it out in the scripts section of package.json.

What is Convex?

Convex is a hosted backend platform with a built-in database that lets you write your database schema and server functions in TypeScript. Server-side database queries automatically cache and subscribe to data, powering a realtime useQuery hook in our React client. There are also Python, Rust, ReactNative, and Node clients, as well as a straightforward HTTP API.

The database support NoSQL-style documents with relationships and custom indexes (including on fields in nested objects).

The query and mutation server functions have transactional, low latency access to the database and leverage our v8 runtime with determinism guardrails to provide the strongest ACID guarantees on the market: immediate consistency, serializable isolation, and automatic conflict resolution via optimistic multi-version concurrency control (OCC / MVCC).

The action server functions have access to external APIs and enable other side-effects and non-determinism in either our optimized v8 runtime or a more flexible node runtime.

Functions can run in the background via scheduling and cron jobs.

Development is cloud-first, with hot reloads for server function editing via the CLI. There is a dashbord UI to browse and edit data, edit environment variables, view logs, run server functions, and more.

There are built-in features for reactive pagination, file storage, reactive search, https endpoints (for webhooks), streaming import/export, and runtime data validation for function arguments and database data.

Everything scales automatically, and it’s free to start.

About

An example of streaming ChatGPT via the OpenAI v4.0 node SDK.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • TypeScript 53.7%
  • CSS 25.6%
  • JavaScript 18.3%
  • HTML 2.4%