Contextualized Information Generation and Retrieval: An Advanced RAG-Based Approach to Natural Language Processing
Generación y Recuperación de Información Contextualizada: Un Enfoque Avanzado Basado en RAG para el Procesamiento del Lenguaje Natural

This work is licensed under a Creative Commons Attribution 4.0 International License.
Esta obra está bajo una licencia internacional
Atribución/Reconocimiento 4.0 Internacional
Show authors biography
This article explores in depth the integration of advanced machine learning techniques using the Retrieval Augmented Generation (RAG) methodology. The dual architecture that combines information retrieval and generation processes is analyzed, highlighting its impact on the training of natural language models. Likewise, specialized variants such as Corrective RAG and Advanced RAG are presented, which incorporate real-time feedback and optimization mechanisms. Also included a mention of the JurislibreIA product, developed by the Sensorama research group, exemplifying practical applications in complex domains such as the legal one. The study is based on implementation examples in Python, explanatory diagrams and a critical review of relevant sources, offering a complete guide for researchers and developers interested in promoting innovative solutions based on RAG.
Article visits 1 | PDF visits 2
Downloads
- P. Lewis, et al., "Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks," Advances in Neural Information Processing Systems (NeurIPS), 2020.
- J. Devlin, M.-W. Chang, K. Lee, and K. Toutanova, "BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding," Proc. of NAACL, 2018.
- T. Mikolov, et al., "Distributed Representations of Words and Phrases and their Compositionality," Advances in Neural Information Processing Systems (NeurIPS), 2013.
- A. Vaswani, et al., "Attention is All You Need," Advances in Neural Information Processing Systems (NeurIPS), 2017.
- I. Sutskever, O. Vinyals, and Q. V. Le, "Sequence to Sequence Learning with Neural Networks," Advances in Neural Information Processing Systems (NeurIPS), 2014.
- S. Hochreiter and J. Schmidhuber, "Long Short-Term Memory," Neural Computation, vol. 9, no. 8, pp. 1735–1780, 1997.
- Y. Kim, "Convolutional Neural Networks for Sentence Classification," Proc. of EMNLP, 2014.
- R. Johnson, et al., "Google's Neural Machine Translation System: Bridging the Gap between Human and Machine Translation," arXiv:1609.08144, 2016.
- Z. Yang, et al., "XLNet: Generalized Autoregressive Pretraining for Language Understanding," Advances in Neural Information Processing Systems (NeurIPS), 2019.
- K. He, X. Zhang, S. Ren, and J. Sun, "Deep Residual Learning for Image Recognition," Proc. of CVPR, 2016.
- M. Abadi, et al., "TensorFlow: A System for Large-Scale Machine Learning," Proc. of OSDI, 2016.
- F. Chollet, "Keras," GitHub repository, 2015. [Online]. Available: https://github.com/keras-team/keras.
- P. Russell and S. Norvig, Artificial Intelligence: A Modern Approach, 3rd ed. Prentice Hall, 2010.
- S. Bengio and Y. LeCun, "Scaling Learning Algorithms towards AI," in Large-Scale Kernel Machines, 2007.
- Hugging Face, "Retrieval-Augmented Generation (RAG)," [Online]. Available: https://huggingface.co/docs/transformers/model_doc/rag Accessed: Feb. 18, 2025.