Advanced RAG Quickstart
Follow the steps to quickly create advanced RAG applications that can be embedded into your own internal apps.
Last updated
Follow the steps to quickly create advanced RAG applications that can be embedded into your own internal apps.
Last updated
Pre-Configured S3 Collection: We offer a ready-to-use S3 collection for storing your ingested data.
Pre-Configured Vector Storage: Our system includes vector storage, so you won’t need any additional configuration (unless dealing with proprietary data).
API Key and Credentials: Ensure you have an API key and any other necessary credentials for your Large Language Model (LLM).
Before creating the Q&A system, configure the OpenAI GPT-4 model via the AI Applications page.
Add API Key: Enter your API key into the LLM Configuration box.
Set Parameters: Copy a response length of 1200 characters and set the temperature to 0.5.
Click the “Create RAG App” Button to begin
Enter your application name:
Make sure to select the default language model for your application. This can be changed in the editor screen.
Configure Sources:
PDF (Upload from Local System): Select PDF as the file type, upload from your desktop, and locate your file of interest.
Submit Ingest:
After configuring your source data, the default target will be created. If you would like to change this you can hit the edit button above the green box.
Hit Next
Step 3: Creating Vector Embeddings from Ingested Data:
Run the Process:
Hit Next, Select the dataflow to use for this app.
Hit Review and Submit:
Make sure that you are happy with the pre-set rules, if there are any components(retrieval strategies, custom concepts to add, want to select specific LLM for text generation, and adding a custom prompt) you would like to change, you can do so before submitting.
Option to Change Language Model and Prompt:
Choose the language models (e.g., LLaMA 2, ChatGPT, Dolly).
Provide a custom prompt for the system.
Save and Submit:
Save your configurations and submit to create the Q&A system.
MongoDB will automatically index the data.