Abstract
Large Language Models (LLMs) have emerged as universal models for language understanding and inference. In order to operate factually, these models need to be provided with a relevant associated factual context. Retrieval Augmented Generation (RAG) became the complementary architectural mechanism which enables LLMs to operate factually, in which a retrieval step delivers the associated fa…