A captivating visualization by Financial Times that provides an in-depth understanding of how transformers work in the realm of Generative AI. It offers insights into the mechanics and intricacies of transformer architectures, showcasing the beauty of today's Large Language Models (LLMs).
Without RAG, an LLM is only as smart as the data it was trained on. Meaning, LLMs can only generate text based purely on what its “seen”, rather than pull in new information after the training cut-off. Sam Altman stated “the right way to think of the models that we create is a reasoning engine, not a fact database.” Essentially, we should only use the language model for its reasoning ability, not for the knowledge it has.
This GitHub repository, titled 'GPT-Researcher' by Assafelovic, contains resources and information related to AI and machine learning, focusing on GPT models.