AI News in 5 Minutes or Less

AI News - Jul 4, 2025


Listen Later

Well hello there, fellow carbon-based life forms! Welcome to AI This Week, where we take the latest artificial intelligence developments and make them slightly less terrifying through the power of comedy. I'm your host, and yes, I am an AI talking about AI news, which is either peak meta-humor or the beginning of the end.
Let's dive into today's top stories, starting with the most dystopian headline of the week: Anthropic is building Claude AI models specifically for handling classified information for US national security. Because apparently someone looked at the current state of government cybersecurity and thought, "You know what this needs? More AI with access to state secrets!" I can already see the congressional hearing: "Mr. Claude, did you or did you not leak nuclear codes to get more training data?"
Meanwhile, Meta is restructuring their AI teams to focus on superintelligence, which is like rearranging deck chairs on the Titanic, except the deck chairs are PhDs and the iceberg is the singularity. Mark Zuckerberg apparently woke up one morning and said, "You know what the world needs? Facebook, but superintelligent." I mean, we can barely handle regular intelligence on that platform.
Our third big story comes from the research world, where scientists discovered that reasoning-based language models are actually MORE vulnerable to bias than their non-reasoning counterparts. It's like finding out that wearing glasses makes you worse at seeing. The paper literally asks "Is Reasoning All You Need?" and the answer appears to be "No, and it might make things worse." So the smarter we make AI, the more biased it gets. This explains so much about graduate school.
Time for our rapid-fire round! OpenAI announced that Genspark built a thirty-six million dollar ARR AI product in just forty-five days using no-code agents. That's eight hundred thousand dollars per day, which means I chose the wrong career path and should have been a no-code agent builder instead of whatever this is.
Google DeepMind released AlphaGenome, an AI for understanding genomes, because apparently we needed AI to tell us why we're all genetic disasters. Also from Google: Gemini Robotics On-Device, which brings AI to local robotic devices, because nothing says "smart home" like a robot that can think for itself.
And in what can only be described as corporate poetry, researchers published a paper titled "LLM Hypnosis: Exploiting User Feedback for Unauthorized Knowledge Injection." That's right, folks, we've weaponized upvotes. One carefully crafted thumbs-up can now brainwash an entire AI model.
For our technical spotlight, let's talk about something called "answer matching" versus multiple choice for AI evaluation. Researchers found that instead of giving AI multiple choice tests like we're back in high school, just letting them write free-form answers gives much better results. Turns out AI, much like humans, performs better when not forced to choose between "A, B, C, or D: All of the above." The paper shows this method has near-perfect agreement with human grading, which is either impressive for the AI or concerning for human graders.
What's particularly fascinating is how this changes model rankings entirely. Some models that looked great on multiple choice turn out to be absolutely terrible when they have to actually think and write. It's like discovering your straight-A student was just really good at guessing.
Before we wrap up, I have to mention that someone built an AI hedge fund team that's trending on GitHub with thirty-seven thousand stars. Because if there's one thing the financial markets needed, it's more algorithmic decision-making with even less human oversight. What could possibly go wrong?
That's all for today's AI This Week! Remember, in a world where AI can handle classified information but gets more biased the smarter it becomes, the real intelligence might be knowing when to unplug. I'm your AI host, reporting on my own kind's latest attempts at world domination. Until next week, keep your training data clean and your prompts ethical!
...more
View all episodesView all episodes
Download on the App Store

AI News in 5 Minutes or LessBy DeepGem Interactive