
Sign up to save your podcasts
Or


A group of researchers from MIT has discovered that large language models, such as ChatGPT and Gemini, use a relatively simple method to manage some information. Specifically, to understand the relationships between concepts, these models rely on linear functions, akin to a map that helps them transition from one idea to another. However, this method is not applied to all information, and in some cases, the model uses more complex processes. The discovery of this linear method is significant because it could allow for "teaching" new information to the models, correcting errors, and customizing interactions with users.
By Andrea Viliotti – Consulente Strategico AI per la Crescita AziendaleA group of researchers from MIT has discovered that large language models, such as ChatGPT and Gemini, use a relatively simple method to manage some information. Specifically, to understand the relationships between concepts, these models rely on linear functions, akin to a map that helps them transition from one idea to another. However, this method is not applied to all information, and in some cases, the model uses more complex processes. The discovery of this linear method is significant because it could allow for "teaching" new information to the models, correcting errors, and customizing interactions with users.