Chatbot using Langchain's model for chatting with your PDF's |Chatbot using Langchain and RAG Advanced with MultiQueryRetrieve and Multivectorstore|
Chatbot where you can chat with your PDF. It consists of a multiretriever and multivector model. When you insert your PDF it will generate a split and a summary of your documents, where in a vectorial base Qdrant will save the complete document, the split and a summary of the document in different collections respectively. When the user inserts the query, the multiquery retriever will create a query adjacent to the original one, having two queries, where they will be used to search for the documentation in the Qdrant summary and split collections. Only two collections were selected for the search due to the limitation of the number of tokens to be passed to the LLM model.
The LLM model used in this project is gemini-1.5-pro.
- pip install requirements.txt
- Insert your credential of Gemini-Pro in the .env file
- Deploy Qdrant using Docker-Compose --> Run image qdrant
- deployment streamlit : streamlit run main.py
Deployment