Sign up to save your podcastsEmail addressPasswordRegisterOrContinue with GoogleAlready have an account? Log in here.
AI on Air brings you the latest news and breakthroughs in artificial intelligence, explained in a way everyone can understand. With AI itself guiding the conversation, we simplify complex topics, fr... more
FAQs about AI on Air:How many episodes does AI on Air have?The podcast currently has 70 episodes available.
October 29, 2024This AI Paper Explores If Human Visual Perception can Help Computer Vision Models Outperform in Generalized TasksRecent research explores the potential of incorporating human visual perception into computer vision models.Researchers at MIT and UC Berkeley suggest that by mimicking human visual processing, particularly the ability to focus on important features and ignore extraneous information, AI models could achieve improved performance on a range of tasks. This approach aligns with other efforts to bridge the gap between seeing and understanding in artificial intelligence, such as the development of the MoAI model. These developments offer promising avenues for creating more robust and adaptable AI systems....more6minPlay
October 28, 2024Microsoft To Launch 'AI Agents' to Help You Handle Routine TasksMicrosoft is actively developing and deploying AI agents, known as Copilot, which are designed to integrate AI into everyday tasks and make it more accessible. This follows a broader trend in the tech industry, as seen in OpenAI’s recent announcements about its own AI agents. Microsoft’s plans emphasize AI at scale, aligning with its ongoing efforts to integrate AI into products like Windows 11 and its existing Copilot service. The news reflects a growing interest in AI agents as a means to enhance user interactions with technology....more5minPlay
October 27, 2024IBM unveils new open source AI ‘Granite 3.0’ models for businessIBM's release of the open-source AI model, Granite 3.0, is a significant development in the field of business AI.This release follows IBM's earlier work on the AI-Hilbert framework, which combines algebraic geometry and mixed-integer optimization for scientific discovery. By making Granite 3.0 open-source, IBM is demonstrating a commitment to making AI more accessible for businesses, mirroring a trend of open-source AI model releases from other tech giants like Google and Meta. These developments reflect the rapid advancements in AI and the increasing focus on creating AI solutions tailored for specific business needs....more5minPlay
October 26, 2024Refined Local Learning Coefficients (rLLCs): A Novel Machine Learning Approach to Understanding the Development of Attention Heads in TransformersRefined Local Learning Coefficients (rLLCs) are a novel machine learning technique that allows researchers to analyze the specific contributions of individual attention heads within transformer models during training. This method offers a more detailed understanding of how these heads evolve and contribute to the model's overall performance. By tracking the development of attention heads over time, rLLCs can pinpoint critical learning phases and potential optimization opportunities within the training process. This new approach offers a deeper understanding of the inner workings of transformer architectures and has the potential to lead to more efficient training strategies and improved performance in future language models....more7minPlay
October 25, 2024This AI Research from Cohere for AI Compares Merging vs Data Mixing as a Recipe for Building High-Performant Aligned LLMsThis research compares two methods for creating powerful and aligned language models: merging and data mixing. Merging, which combines pre-trained models, outperforms data mixing in terms of both performance and alignment. This suggests that merging is a promising approach for efficiently building more capable and aligned AI systems. The findings are supported by other research exploring the benefits of combining diverse language models....more3minPlay
October 24, 2024Are Brains and AI Converging?—an excerpt from ‘ChatGPT and the Future of AI: The Deep Language Revolution’The provided episode highlights the burgeoning convergence of artificial intelligence (AI) and neuroscience, exploring the ways in which AI is mimicking human cognitive processes. It specifically focuses on the development of large language models like ChatGPT and the MoAI model, which bridges the gap between visual perception and understanding. The episode further suggests that AI is starting to learn language similarly to children, underscoring the growing similarity between human and AI cognitive abilities....more5minPlay
October 23, 2024CREAM: A New Self-Rewarding Method that Allows the Model to Learn more Selectively and Emphasize on Reliable Preference DataCREAM, which stands for Confidence-based REward Adjustment Method, is a new technique for training language models that focuses on improving their performance and alignment by using the model's confidence in its judgments to adjust rewards. This method prioritizes high-confidence preferences while downplaying those with lower confidence, leading to more selective and efficient learning. CREAM builds upon earlier self-rewarding methods, such as those discussed in the Meta and NYU paper on self-rewarding language models and the meta-rewarding technique, by incorporating confidence-based reward adjustments. This approach offers a more refined way to improve AI models through self-improvement and alignment....more6minPlay
October 22, 2024Embed-then-Regress: A Versatile Machine Learning Approach for Bayesian Optimization Using String-Based In-Context RegressionThis episode explores a novel method for enhancing large language models (LLMs) through "self-reflection." Researchers have devised a technique that allows LLMs to analyze and predict their own behavior, resulting in improved accuracy and reliability. This approach, achieved by fine-tuning LLMs on datasets containing both correct and incorrect responses alongside explanations, fosters increased transparency and trust in AI systems. By enabling LLMs to generate explanations and anticipate errors, this method contributes significantly to the development of more self-aware and reliable AI technologies....more3minPlay
October 21, 2024Graph-Constrained Reasoning (GCR): A Novel AI Framework that Bridges Structured Knowledge in Knowledge Graphs with Unstructured Reasoning in LLMsGraph-Constrained Reasoning (GCR) is a new AI framework that integrates knowledge graphs with large language models (LLMs). This approach aims to improve the accuracy and reliability of AI systems by utilizing the strengths of both structured and unstructured reasoning. GCR bridges the gap between these two types of knowledge representation, enhancing the capabilities of LLMs by using knowledge graphs to guide and constrain their outputs. This framework aligns with other recent AI developments, such as CodexGraph and GPTSwarm, which all aim to build more robust and versatile AI systems by combining multiple techniques....more6minPlay
October 20, 2024MMed-RAG: A Versatile Multimodal Retrieval-Augmented Generation System Transforming Factual Accuracy in Medical Vision-Language Models Across Multiple DomainsNvidia AI has developed a new type of transformer architecture called the Normalized Transformer (nGPT) that uses hypersphere-based normalization. This innovation significantly speeds up the training process for large language models (LLMs) by 4 to 20 times while also improving stability. The nGPT addresses the challenges of vanishing gradients and unstable training dynamics present in traditional transformer models. By restricting activations to a hypersphere, it creates a more stable training landscape, leading to faster convergence and better performance. This breakthrough has the potential to accelerate LLM training and reduce computational costs....more5minPlay
FAQs about AI on Air:How many episodes does AI on Air have?The podcast currently has 70 episodes available.