Skip to content

An LLM-powered chat application that uses Django, React, and LangChain.

Notifications You must be signed in to change notification settings

virattt/chat_app

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

55 Commits
 
 
 
 
 
 
 
 

Repository files navigation

LLM-Powered Chat Application

This repository contains all of the starter code needed to run an LLM-powered chat app on your local machine:

  1. Django backend
  2. React TypeScript frontend
  3. LangChain Agents and LLMs

Getting Started 🚀

To run the chat app, you need to:

  1. Clone this GitHub repo
  2. Run the backend server
  3. Run the frontend app

1. Clone this GitHub repo 📁

To clone this GitHub repo, open up your Terminal (MacOS) or Bash terminal (Windows) and navigate to wherever you want to save this repo on your local machine. Then, run:

git clone https://github.com/virattt/chat_app.git

Make sure that you have git installed (instructions).

2. Run the backend server 🖥️

Once you have this chat_app project cloned locally, navigate to the backend directory:

cd ~/path_to/chat_app/backend

Create and activate a virtual environment:

python3 -m venv myenv

For MacOS/Linux:

source myenv/bin/activate

For Windows:

myenv\Scripts\activate

Install the necessary libraries:

pip install -r requirements.txt

Make sure that you have Redis installed. You can find instructions here. Once installed, run redis:

redis-server

Run the backend server:

daphne project.asgi:application

If your backend server is running correctly, you should see something like this:

"WSCONNECTING /ws/chat/" - -
"WSCONNECT /ws/chat/" - -

Important: In order to run the LLM, set your Open AI API key here.

3. Run the frontend app 💻

In a new Terminal window (or tab), navigate to the frontend directory:

cd ~/path_to/chat_app/frontend

Make sure that you have Node and npm installed (MacOS instructions and Windows instructions)

Install the necessary packages:

npm install

Run the frontend app:

npm start

If successful, your browser should open and navigate to https://localhost:3000/. The chat app should load automatically.

The Chat App UX 🤖

As of May 17, 2023 Screen Shot 2023-05-17 at 4 52 27 PM

Troubleshooting ⚠️

If you encounter any issues, send me a message on Twitter!

About

An LLM-powered chat application that uses Django, React, and LangChain.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published