Systems Thinking and Beyond

Understanding Large Language Model AIs


Listen Later

The AI team takes a deep dive into the technical architecture and operational logic of Large Language Models (LLMs). They explain that these systems are trained through a multi-stage process; pre-training, fine-tuning, and human feedback, to predict text sequence. A central focus is the Transformer architecture, which uses an attention mechanism to understand relationships between words and manage linguistic nuances such as spelling errors. The team clarify that AI "memory" is actually a process where the entire conversation history is re-read during every interaction to maintain coherence. Finally, the team define LLMs as probabilistic state machines that, despite their sophisticated conversational abilities, face limitations such as factual hallucinations and fixed knowledge cutoffs.

...more
View all episodesView all episodes
Download on the App Store

Systems Thinking and BeyondBy Dr Joseph Kasser