BastinFlorian / gcp-llm-retrieval-augmentation

A retrieval augmentation LLM demo in GCP

Home Page:https://llmops-demos-frg.web.app

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

LLM retrieval augmentation in Google Cloud

This demo features GCP Matching Engine and VertexAI PaLM to combine the functionality of retrieval augmentation and conversational engines to create a question answering system where the user can ask a question and the LLM will use it's given context to answer the question.

The Dataset used is the Stanford Question Answering Dataset (SQuAD) , a reading comprehension dataset, consisting of questions posed by crowdworkers on a set of Wikipedia articles.

The demo can be accessed here.

Services used

Architecture

Frameworks:

Prerequisites

Docs

  1. Infrastructure and Matching Engine Setup: Setup the required infrastructure using Terraform and create the Matching Engine index
  2. Create embeddings: Generate the embeddings for the documents and index them in Matching Engine
  3. Firestore: Index the documents in Firestore
  4. LangChain Retriever and Agent: Create a LangChain retriever and conversational agent
  5. Cloud Run: Grab all the code, package it and deploy the API to Cloud Run
  6. Firebase WebUI: Create the Web app

About

A retrieval augmentation LLM demo in GCP

https://llmops-demos-frg.web.app


Languages

Language:Jupyter Notebook 68.3%Language:Svelte 14.7%Language:HCL 7.3%Language:Python 6.4%Language:JavaScript 1.1%Language:HTML 0.7%Language:CSS 0.5%Language:TypeScript 0.5%Language:Dockerfile 0.4%