How to Face Developer to Well Questions Answering Questions

Engineering is one of the most relevant topics for a machine learning today, which is why I write my third article on the topic. My goal in both of my understanding goals for the engineering conditions of llms and shares that information using my articles.
In today's article, I will discuss the luxury context to answer the question. Usually, this context is based on the restoration of unpleasant generations (RAG), however, today's variable area, this method should be renewed.
You can also read previous engineering ethics:
- Employment Basic Engineering Techniques
- Competition strategies have improved
Content
Why should you care for the engineering of the bounds
First, let me reveal three important points of why you have to take care of the engineering of the Bocho:
- Better Issue format By avoiding the context. Few tokens are not required to increase the quality of the output. You can read more information on this article
- Inexpensive (Do not send unnecessary tokens, they charge money)
- Speed (small tokens = fast response times)
These are three key artrics of many programs to answer questions. The quality of the most important nature of nature, processing users will not want to use the program.
In addition, the price should be considered, and if you can reduce (other than multi-engineering costs), a simple decision to do so. Finally, the quick question response process provides a better user experience. You don't want users to wait a lot of seconds to get an answer when Chatgpt will respond as soon as possible.
How to Answer Traditional Questions
In this sense, this general question of responding to the programs built after ChatGPT release. This program is traditional rag, which works as follows:
- Download Documented Documents in the User Question, using the resume of the Vector Same
- Feed the relevant documents and question to the llm, and you have received feedback
Thinking about its simplicity, this method works best. It's interesting enough, we see this also in yet another traditional way. BM25 has been available since 1994, for example, it was later used by anthropic when they presented practical returns, proving that effective return techniques.
However, you can still enhance your application for answering the question by renewing your rag using other strategies I will describe us in the next section.
To improve downloading the rag status
Or the RAG works well, you may have achieved the better performance by introducing the strategies to discuss in this section. The techniques that I explain here is all focused on improving the context that is llm. You can improve this context in two advanced ways:
- Use a few tokens in the wrong context (for example, to remove or use small assets from the relevant documents)
- Add relevant documents
Therefore, you should focus on achieving one of the above point points. If you think of the terms of accuracy and remembering:
- Increases the accuracy (at the memory cost)
- Increase to remember (at cost of accuracy)
This trading should do while working with Excusp System Engineering System.
Reducing the amount of unfair tokens
In this section, I highlight three key ways to reduce the number of unfair token in the context of the City:
- Reranking
- Condensation
- Promotes GPT
When managing documents from the Vector's similarity searches, restored in the correct order should be given, given points like Vector. However, this Brack Skiness may not appear properly what relevant documents are.
Reranking
That way you can use a Reward model, for example, QWEN Ranker, editing document chunks. You can only choose to save high chunks of X relevant (according to Reranker), which should remove the unsupervised text from your context.
Condensation
You can also select summarizes documents, reduce the number of tokens used for each document. For example, you can save a complete document from 10 high tracked documents, summarizing the documents set from 11-20, drop left.
This method will increase the risk of keeping the complete context from the relevant documents, and at least keep the context from the information options.
Promotes GPT
Finally, don't again raise GPT if the specified documents are related to the user's question. For example, if you download 15 documents, you can make a 15 llm able to judge if each document is appropriate. You throw documents considered inappropriate. Remember that these llm calls need to be scattered to save time to answer within the acceptable limit.
Adding relevant documents
Before or after removing improper texts, and confirms that you are installing the appropriate documents. Including two main ways in this paragraph:
- The models are better embedded
- Search for multiple documents (at the low cost of the lower case)
The models are better embedded
Finding the best incentive models, you can go to the leading model, where Gemini and QWen is high quality 3 as the writing of this article. Updating your Mending Model is usually the cheap way to download the correct documents. This is because the running and maintenance is often cheaper, for example, Gemino API, and the maintenance of PineCones.
Search other texts
Another way (simple) How to download the appropriate documents to download other texts usually. Downloading multiple documents naturally increases the possibilities that you add stakeholders. However, you must measure this by avoiding the context and reduces the number of inappropriate documents. All unnecessary tokens in the llM call, almost before, it is possible:
- Reduce the quality of output
- Increase costs
- Low speed
These are the key features of the application program.
Search Method at Agentic
I discuss ways of Agentic Search ways in the topics previous, for example, when I discuss your AI search. However, in this stage, I will wash deep in search of a powerful search, which takes a specific location or the action to return the vector to your RAG.
The first step is that the user provides their question about a given set of data points, for example, a set of documents. Then sets the Ageentic program containing the orchestra for the agent and the list of underlying.
This is an example of the pipe agents that will follow (although there are many setup methods).
- The Orchestra Agent tells two sugars to benefit from every word of the documentation and restore the relevant documents
- Due documents are taught in orchestra for the agent, and also dispose of any relevant documentation, downloading Subparts (Chunks) that corresponds to the user's question. These fed chunks returned to orchestra
- The Orchestra Agent Answers User Question, Given the Tributed Chunks
Another flow you can use can keep the document embark on, and replace the first step in the vector between the user question and each document.
This Agentic method has upsudes and decreases.
Upsides:
- A better chance to download chunks worth the chunks than with traditional rag
- More control of the RAG system. You can update the system's motivation, etc. when the RAG is basic the same as its embarking
Down:
In my opinion, create a return program for Agen Agent a powerful force that can lead to incredible effects. The consideration to do when creating such a system is whether more quality is (possible) may not qualify for costs increase.
Some of the characteristics of context
In this article, I have greatly covered the Scriptural engineering engineering to collect documents in response to a question response program. However, there are other factors to monitor, especially:
- Program / user you use
- Some information raped raped
Prectu you write about your program of response system should be accurate, organized, and avoid inappropriate information. You can read many other articles on the topic of the shape encouragement, and you can usually ask the llm to improve these features of your Prompt.
Sometimes, feeding other details as quickly as you can. A common example feeding on Metadata, for example, data covering information about the user, such as:
- Name
- Growth
- Which usually searches about
- entll
Whenever you add such information, you must always:
Does the amendment of this information help my program response question answer the question?
Sometimes the answer is yes, sometimes not. The most important part is to make a reasonable decision that the knowledge is needed. If you can't be infected with this information where it is fast, it should be deleted.
Store
In this article, I discussed the Kingdom engineer about your question response program, and why it is important. Question response programs often have the first step to download the relevant information. Focusing on this information should reduce the amount of unfair tokens, while including as many pieces of information as possible.
👉 I have found in the community:
🧑💻 Contact your
🔗 LickDin
🐦 x / Twitter
✍️ Medium
You can also read my deeper article in the restoration of anthropic content below:



