commercial value of LLM in knowledge-intensive 
industries in the near future. Its ability to solve model 
illusion, update model information base in real time 
and keep it private all show its potential to improve 
the accuracy, reliability and stability of question 
answering field, and make the generated text results 
more realistic. These will lead to it becoming an 
inseparable part of the future private deployment of ai 
products, such as enterprise private document review, 
industry bid assistance writing and other customized 
functions. The great promise of the business sector 
also means that more vertical-specific assessment 
methods or test datasets is required. At the same time, 
the dynamic information field such as finance and 
news media to establish a regular data update process 
to meet the needs of the industry. This process can 
automatically complete the extraction, analysis and 
update of new data to meet the needs of the industry. 
RAG enhanced retrieval technology also 
encounters limitations. For example, traditional 
vector retrieval cannot represent logical reasoning 
connections due to its embeddedness and lacks real 
relevance and thought chain. The deficiency of 
knowledge base context and the loss of key 
information in the process of compressing paragraph 
vector into single vector will inevitably lead to the 
problem of knowledge waste. Who will dominate 
RAG Retrieval and Generation more, and whether 
their performance in different fields will make their 
emphases different, these questions will directly lead 
to whether rag will be oriented towards search or 
agent in the future technological development path. 
The constraints or balance points between these 
should be focused, these are still unsolved challenges. 
Challenges come with new technological 
possibilities. For example, the use of knowledge 
graph embedding makes the generated results more 
interpretable and logical upward compatible, so that 
logical reasoning is more accurate. For the pain points 
of insufficient context in the knowledge base, the 
enhancement of context information by using more 
efficient document parsing tools can be ensured to 
add relevant metadata to each paragraph of text. 
4  CONCLUSION 
This paper mainly summarizes RAG's recent 
algorithms for improving retrieval efficiency and the 
impact of weak interference brought by improving 
user prompts on generated results, as well as the 
introduction and discussion of vector data reranking 
and truncation strategies. The current RAG 
enhancement was introduced to enhance the retrieval 
robustness and enhance the explainability of the 
query method. The RAG application in the current 
medical question answering system is introduced, and 
the performance difference of its hybrid training 
method in the application side is presented. By using 
the evaluation tool suitable for the field of medical 
question answering, the combination of LLM and 
retrieval device suitable for this field is obtained. 
Finally, the commercial application prospect and 
future research direction of RAG are discussed. There 
is reason to believe that RAG enhanced retrieval 
technology will become an important part of the 
privatization ai deployment boom in the future. 
REFERENCES 
Chang, Y., Wang, X., Wang, J., Wu, Y., Yang, L., Zhu, K., 
... & Xie, X. 2023. A survey on evaluation of large 
language models. ACM Transactions on Intelligent 
Systems and Technology. 
Chen, J., Lin, H., Han, X., & Sun, L. 2024. Benchmarking 
large language models in retrieval-augmented 
generation. In Proceedings of the AAAI Conference on 
Artificial Intelligence (Vol. 38, No. 16, pp. 17754-
17762). 
Hu, Z., Wang, C., Shu, Y., & Zhu, L. 2024. Prompt 
perturbation in retrieval-augmented generation based 
large language models. arXiv preprint 
arXiv:2402.07179. 
Kasneci, E., Seßler, K., Küchemann, S., Bannert, M., 
Dementieva, D., Fischer, F., ... & Kasneci, G. 2023. 
ChatGPT for good? On opportunities and challenges of 
large language models for education. Learning  and 
individual differences, 103, 102274. 
Kenneweg, T., Kenneweg, P., & Hammer, B. 2024. 
Retrieval Augmented Generation Systems: Automatic 
Dataset Creation, Evaluation and Boolean Agent 
Setup. arXiv preprint arXiv:2403.00820. 
Lewis, P., Perez, E., Piktus, A., Petroni, F., Karpukhin, V., 
Goyal, N., ... & Kiela, D. 2020. Retrieval-augmented 
generation for knowledge-intensive nlp tasks. 
Advances in Neural Information Processing 
Systems, 33, 9459-9474. 
Xiong, G., Jin, Q., Lu, Z., & Zhang, A. 2024. 
Benchmarking retrieval-augmented generation for 
medicine. arXiv preprint arXiv:2402.13178. 
Xu, S., Pang, L., Xu, J., Shen, H., & Cheng, X. 2024. List-
aware reranking-truncation joint model for search and 
retrieval-augmented generation. arXiv preprint 
arXiv:2402.02764. 
Zhao, P., Zhang, H., Yu, Q., Wang, Z., Geng, Y., Fu, F., ... 
& Cui, B. 2024. Retrieval-Augmented Generation for 
AI-Generated Content: A Survey. arXiv preprint 
arXiv:2402.19473. 
Zhao, W. X., Zhou, K., Li, J., Tang, T., Wang, X., Hou, Y., 
... & Wen, J. R. 2023. A survey of large language 
models. arXiv preprint arXiv:2303.18223.