Skip to content

Enhance LLM retrieval performance with Azure Cosmos DB Semantic Cache. Learn how to integrate and optimize caching strategies in real-world web applications.

License

Notifications You must be signed in to change notification settings

jonathanscholtes/LLM-Performance-with-Azure-Cosmos-DB-Semantic-Cache

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

3 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Azure Cosmos DB Semantic Cache - LangChain Q&A RAG

Overview

This repository contains a demos showcasing the implementation of the RAG (Retrieval Augmented Generation) pattern using Azure Cosmos DB for MongoDB vCore with Semantic Cache and LangChain. The RAG pattern combines retrieval-based and generative-based approaches to natural language processing, enhancing text generation capabilities.

Features

  • Integration of Azure Cosmos DB for MongoDB vCore as a scalable and fully managed database solution.
  • Utilization of LangChain for text processing, enabling retrieval of relevant information and generation of contextually relevant responses.
  • Demonstrates how to implement the RAG pattern for enhanced natural language processing tasks.

Requirements

  • Azure subscription for deploying Azure Cosmos DB for MongoDB vCore.
  • Python environment with LangChain installed.
  • Basic knowledge of MongoDB and natural language processing concepts.

Usage

  1. Follow the steps provided in the README file.

Steps

  1. Step 1 - Load Cosmos DB for Mongo DB Vector Store using sample dataset
  2. Step 2 - Create FastAPI to integrate LangChain RAG pattern with web front-end.
  3. Step 3 - Build the React web front-end to ask 'grounded' questions of your data and view relevant documents.
  4. Follow the setup instructions provided in the README file.
  5. Run the demo application and explore the RAG pattern in action.

License

This project is licensed under the MIT License, granting permission for commercial and non-commercial use with proper attribution.

Support

For any questions or issues, please open an issue on GitHub or reach out to the project maintainers.

Disclaimer

This demo application is provided for educational and demonstration purposes only. Use at your own risk.

About

Enhance LLM retrieval performance with Azure Cosmos DB Semantic Cache. Learn how to integrate and optimize caching strategies in real-world web applications.

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published