Skip to content

QnA application made using OpenAI API, Redis and Docker

License

Notifications You must be signed in to change notification settings

TheCleverIdiott/QnA

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

7 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Question & Answering using Redis & OpenAI

Redis plays a crucial role in the LLM & GenAI wave with it's ability to store, retrieve, and search with vector spaces in a low-latency, high-availability setting. With its heritage in enterprise caching, Redis has both the developer community and enterprise-readiness required to deploy quality AI-enabled applications in this demanding marketplace.

OpenAI is shaping the future of next-gen apps through it's release of powerful natural language and computer vision models that are used in a variety of downstream tasks.

This example Streamlit app gives you the tools to get up and running with Redis as a vector database and OpenAI as a LLM provider for embedding creation and text generation. The combination of the two is where the magic lies.

ref arch


Run the App

Create your env file:

$ cp .env.template .env

fill out values, most importantly, your OPENAI_API_KEY.

Run with docker compose:

$ docker compose up

add -d option to daemonize the processes to the background if you wish.

Navigate to:

http://localhost:8080/

The first time you run the app -- all documents will be downloaded, processed, and stored in Redis. This will take a few minutes to spin up initially. From that point forward, the app should be quicker to load.

Ask the app anything about the 2020 Summer Olympics...

About

QnA application made using OpenAI API, Redis and Docker

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published