Skip to content

feat: Integrate Milvus, Ollama, PDF ingestion, and Streamlit UI #439

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Open
wants to merge 1 commit into
base: master
Choose a base branch
from

Conversation

Sasapu-Hemasai
Copy link

Key changes include:

  • Replaced Weaviate with Milvus as the vector database.
  • Added support for Ollama for both embeddings and language model generation, configurable alongside existing OpenAI integrations.
  • Implemented ingestion pipeline to process PDF documents from a local knowledge_docs folder. Includes text extraction and metadata assignment. A placeholder is included for your logic to extract source URLs from PDF content for citations.
  • Developed a Streamlit-based frontend for your interaction, allowing model selection and displaying chat history with sources.
  • Updated the backend server (backend/server.py) to correctly serve the LangServe graph using FastAPI.
  • Provided comprehensive README.md documentation covering setup, configuration, data preparation (including the PDF URL extraction task), and instructions for running all components.
  • Removed Weaviate dependencies and updated project configurations.

Key changes include:
- Replaced Weaviate with Milvus as the vector database.
- Added support for Ollama for both embeddings and language model generation,
  configurable alongside existing OpenAI integrations.
- Implemented ingestion pipeline to process PDF documents from a local
  `knowledge_docs` folder. Includes text extraction and metadata assignment.
  A placeholder is included for your logic to extract source URLs
  from PDF content for citations.
- Developed a Streamlit-based frontend for your interaction, allowing
  model selection and displaying chat history with sources.
- Updated the backend server (`backend/server.py`) to correctly serve
  the LangServe graph using FastAPI.
- Provided comprehensive `README.md` documentation covering setup,
  configuration, data preparation (including the PDF URL extraction task),
  and instructions for running all components.
- Removed Weaviate dependencies and updated project configurations.
Copy link

vercel bot commented May 23, 2025

@google-labs-jules[bot] is attempting to deploy a commit to the LangChain Team on Vercel.

A member of the Team first needs to authorize it.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

1 participant