This is a repo for a human-interface LLM-powered QA Slack chatbot for answering questions related to DataTalks.Club Zoomcamps
Please follow this report that explains in detail how the bot works
Course FAQ Google Document and the Course repo get indexed to the Pinecone vector store. Then semantic search retrieves the most similar (and hopefully most relevant) pieces to the question asked. Then this information is passed as a context to a conversational LLM to form the final answer.
Use any python virtual environment manager of your preference and install the dependencies from requirements.txt
Fill Pinecone index
-
Set
PINECONE_API_KEY
andPINECONE_ENV
environmental variables accordingly -
run ingest/ingest.py
python ingest_mlops_old.py
-
Set
OPENAI_API_KEY
,PINECONE_API_KEY
, andPINECONE_ENV
environmental variables accordingly -
Run test.py
python test.py
- Set
OPENAI_API_KEY
,SLACK_APP_TOKEN
,SLACK_BOT_TOKEN
,PINECONE_API_KEY
, andPINECONE_ENV
environmental variables accordingly - To launch the bot on the CLI run slack_bot/main.py
python main.py
Alternatively it can be launched with Docker, please follow this README