This blueprint allows is a one click to deploy a RAG Slim pipeline for inference using LLM connected to cnvrg storage solution.
- A Large Language Model hosted on cnvrg, OpenAI or HuggingFace.
- A cnvrg dataset holding the relevant documents to be used for RAG endpoint. The dataset needs to be added to the flow as a data task.
- In order to keep the FastRAG endpoint up-to-date with newly added data use the continual learning feature in the flow configurations. For every file change of the conected dataset a new version of the FastRAG endpoint will be launched with access to the latest files.
- Click on
Use Blueprintbutton. - You will be redirected to a new project with the blueprint flow page.
- Go to the project settings section and update the environment variables with relevant information that will be used by the RAG endpoint. For more info see the component documentation
- Link the cnvrg dataset as a task with the inference.
- Click on continual learning and select
Trigger on dataset updateand choose your dataset - Click on the ‘Run Flow’ button
- In a few minutes you will have a RAG endpoint
- Go to the ‘Serving’ tab in the project and look for your endpoint.
- You can use the Try it Live section to query the RAG endpoint and generate relevant answers with LLM connected.
- You can also integrate your API with your code using the integration panel at the bottom of the page