
Photo by the author Canva
We all know the two main problems, which have been indicated as the main disadvantages of immense language models (LLM):
- Hallucinations
- No updated information except for their knowledge of knowledge
Both of these issues raised grave doubts as to the credibility of LLM results, and there the generation of download (RAG) appeared as a powerful way to solve them, offering more correct, conscious contextual answers. Nowadays, it is widely used in various industries. However, many beginners got stuck in the study of only one elementary architecture: basic search for vectors by text documents. Sure, it works for most basic needs, but limits creativity and understanding.
This article has a different approach. Instead of a deep immersion in one narrow configuration to explain the details of one rag (Like advanced hints, fragment, deposition and recovery), I think that beginners apply more from studying a wide spectrum of rag patterns. In this way you will see how adaptable and versatile the concept of rag is really and get inspired to create your own unique projects. Let’s look at five entertaining and engaging projects that I have prepared that will facilitate you do it. Let’s start!
# 1
To combine: https://www.youtube.com/watch?v=hrvyei7vfsm

Start from scratch with the construction of a elementary RAG system. This beginner project shows how to build a RAG system that answers questions from any PDF file using an open source model, such as Lama2 without paid API interfaces. You will start Lama2 locally with OllamLoad and divide PDF with PyPDF With LangchainCreate deposition and store them in a vector shop in memory Docaray. Then you will configure the recovery chain Langchain To download the appropriate fragments and generate answers. Along the way, you will learn the basics of working with local models, pipeline collection buildings and testing results. The end result is a elementary bot, which can answer PDF specific questions, such as “What is the cost of the course?” with an correct context.
# 2. Multimodal rag: conversation with PDF containing paintings and tables
To combine: https://youtu.be/ulrreyh5c0?feature=shared

In the previous project we only worked with text data. Now it’s time to even out. Multimodal RAG expands established systems to process image, tables and text in PDF files. In this tutorial Alejandro Ao goes through the apply of such tools Langchain and Unstructured Library for processing mixed content and transmitting it into multimodal LLM (e.g. GPT-4 with vision). You will learn how to extract and embed text, images and tables, combine them into a uniform prompt and generates answers that understand the context in all formats. Built -in will be stored in a vector database and Langchain The recovery chain will connect everything so that you can ask questions such as “Explain the chart on page 5.”
# 3. Creating a rag on a device with Objectbox and Langchain
To combine: https://www.youtube.com/watch?v=9lewl1bus6g

Now let’s go fully locally. This project will lead you through the construction of a rag system, which operates completely on the device (without a cloud, without the Internet). In this tutorial you will learn how to store data and set locally with a delicate, ultra-expensive Objectbox Vector database. You will apply Langchain To build a download and generation pipeline so that your model can answer questions from documents directly on the computer. It is ideal for everyone interested in privacy, data control or simply willingness to avoid the cost of the API interface. In the end you will have a system of AI questions and answers, which lives on your device, reacting quickly and safely.
# 4. Building a real -time pipeline with Neo4J and Langchain
To combine: https://www.youtube.com/watch?v=ik8Gnjj-13i

In this project you will go from ordinary documents to powerful charts. This tutorial shows how to build a RAG system in real time using the facilities of the knowledge chart. You will work in a notebook (like Colab), configure Neo4J Instance in the cloud and create nodes and edges to represent your data. Then using LangchainYou will connect your LLM chart to generate and download, allowing you to ask for contextual relationships and visualize the results. It’s a great way to learn the logic of the chart, Cypher Query and method of merging structured knowledge of the chart from Clever Ai Answers. I also wrote a thorough guide on this subject, Building a rag system: step by step approachWhere I break down, how to create a Graphr configuration from scratch. Check it out if you prefer tutorials based on the article.
# 5. Implementation of an agent rag with the Llam index
To combine: https://youtube.com/playlist?list=PLU7AW4OZXRJADVRIADBAMASWFALA0&feature=shared

In earlier projects we focused on downloading and generating, but here the goal is to make RAG “Agentic” by giving her reasoning of loops and tools so that it can solve problems in many steps. This list of playback of Prince Kramph is divided into 4 stages:
- Router query engine: Configure Lama-Index To direct questions to the right source, such as the VS. Vector Index. Summary index
- Connection of functions: Add tools such as API calculators or interfaces so that the rag can download live data or perform tasks in flight
- Multi -stage reasoning: Break up complicated queries into smaller subfamily (“Summary first and then analyze”)
- Over many documents: Scale your reasoning in several documents simultaneously with agents serving power
# Wrapping
And here you have: 5 -friendly RAG projects that go beyond the usual configuration of “vector text search”. My advice? Do not strive for perfection during the first attempt. Choose one project, herring and let yourself experiment. The more designs you study, the easier it will be to mix and match ideas to your own custom RAG applications. Remember that real fun begins when you stop simply “recovering” and start “thinking” about how your artificial intelligence can reason, adapt and interact in a smarter way.
Canwal Mehreen He is a machine learning engineer and a technical writer with a deep passion for data learning and AI intersection with medicine. He is the co -author of the ebook “maximizing performance from chatgpt”. As a Google 2022 generation scholar for APAC, it tells diversity and academic perfection. It is also recognized as a variety of terradate at Tech Scholar, Mitacs Globalink Research Scholar and Harvard Wecode Scholar. Kanwalwal is a warm supporter of changes, after establishing FemCodes to strengthen women in the STEM fields.
