A cloud-native vector database, storage for next generation AI applications
-
Updated
Dec 20, 2024 - Go
A cloud-native vector database, storage for next generation AI applications
Chat and Ask on your own data. Accelerator to quickly upload your own enterprise data and use OpenAI services to chat to that uploaded data and ask questions
RAFT contains fundamental widely-used algorithms and primitives for machine learning and information retrieval. The algorithms are CUDA-accelerated and form building blocks for more easily writing high performance applications.
cuVS - a library for vector search and clustering on the GPU
Harnessing the Memory Power of the Camelids
Semantic product search on Databricks
🗲 A high-performance on-disk dictionary.
just testing langchain with llama cpp documents embeddings
Multi-User Chatbot with Langchain and Pinecone in Next.JS
The official Elasticsearch .NET Vector Store Connector for Microsoft Semantic Kernel.
A command-line tool that ingests documents and generates instant answers to your questions about those documents using ChatGPT, giving you the Sheldon Cooper you never had at your fingertips.
The AI Assistant uses OpenAI's GPT models and Langchain for agent management and memory handling. With a Streamlit interface, it offers interactive responses and supports efficient document search with FAISS. Users can upload and search pdf, docx, and txt files, making it a versatile tool for answering questions and retrieving content.
A Question Generation Application leveraging RAG and Weaviate vector store to be able to retrieve relative contexts and generate a more useful answer-aware questions
A simple way to convert and manage files in vector storage.
AI Data Assistant empowers data professionals by speaking both human language and machine langauges (SQL/Python)
AgriGenius: AI-Powered Agriculture Chatbot is a Python web application designed to empower farmers with information accessibility. AgriGenius leverages a Retrieval-Augmented Generation model to address farmer's agricultural queries with precise answers.
This is a toy Web application written in Flask, featuring a Medical Assistant Chatbot powered by Large Language Models (LLMs) and Retrieval Augmented Generation (RAG).
minimem is a minimal implementation of in-memory vector-store using only numpy
Add a description, image, and links to the vector-store topic page so that developers can more easily learn about it.
To associate your repository with the vector-store topic, visit your repo's landing page and select "manage topics."