Kirsten Odendaal

Context Retrival Language Model: Creation of a RAG Demonstrator

RAG LLM

Challenges of Large Language Models

Large Language Models (LLMs) like ChatGPT have significantly advanced natural language processing by enabling machines to understand and generate human-like text. However, developing and deploying these models present several challenges:


General Cost of Building an LLM

Training LLMs is resource-intensive, illustrated by the GPU hours required:

To provide context, ChatGPT-3.5 has 175 billion parameters, and ChatGPT-4.0 is estimated to have 1.76 trillion parameters.


Proposed Solutions

Given the high costs and complexity of building LLMs from scratch, leveraging existing models is a more feasible approach. This can be achieved through:


Retrieval-Augmented Generation (RAG)

One form of context injection is known as Retrieval-Augmented Generation (RAG). RAG combines LLMs with a retrieval system to improve performance and address key challenges:

rag_framework


Applications

By automating mundane tasks, LLMs free up valuable time for creative and strategic activities, increasing overall productivity and efficiency.


Github Repository

Github Repository Link