Welcome to The Nonlinear Library, where we use Text-to-Speech software to convert the best writing from the Rationalist and EA communities into audio. This is: AI #19: Hofstadter, Sutskever, Leike, published by Zvi on July 6, 2023 on LessWrong.
The big news of the week is OpenAI introducing the Superalignment Taskforce. OpenAI is pledging 20% of compute secured to date towards a four year effort to solve the core technical challenges of superintelligence alignment. They plan to share the results broadly. Co-founder and OpenAI Chief Scientist Ilya Sutskever will lead the effort alongside Head of Alignment Jan Leike, and they are hiring to fill out the team.
That is serious firepower. The devil, as always, is in the details. Will this be an alignment effort that can work, an alignment effort that cannot possibly work, an initial doomed effort capable of pivoting, or a capabilities push in alignment clothing?
It is hard to know. I will be covering these questions in a post soon, and want to take the time to process and get it right. In the meantime, the weekly post covers the many other things that happened in the past seven days.
Table of Contents
Introduction.
Table of Contents.
Language Models Offer Mundane Utility. Commercial LLMs keep the edge.
Language Models Don’t Offer Mundane Utility. Can you?
Fun With Image Generation. Valve says: Keep all that fun away from Steam.
They Took Our Jobs. Then they came for America’s finest news source.
Introducing. If only it were as easy to forget.
Reinforcement Learning By Humans From Feedback. United we grok.
Costs are not Benefits. Easy mistake to make. Many such cases.
Quiet Speculations. Questions that are not so infrequently asked.
The Quest for Sane Regulation. Some small good signs.
Another Open Letter. EU CEOs call on EU to not act like the EU.
The Week in Audio. Odd Lots offers AI fun.
No One Would Ever Be So Stupid As To. Is that what you think?
Safely Aligning a Smarter than Human AI is Difficult. Formal verification?
Rhetorical Innovation. The latest crop of explanations and resources.
People Are Worried About AI Killing Everyone. Including Douglas Hofstadter.
The Lighter Side. Actual progress.
Language Models Offer Mundane Utility
Open source models often do well on open source benchmarks. So do foreign efforts like Falcon or Ernie. When it comes to actual mundane utility, or tests that were not anticipated in advance, the answer seems to come back somewhat differently.
Lmsys.org: Quick note - we’ve transitioned from the deprecated vicuna benchmark to a more advanced MT-bench, including more challenging tasks and addressing biases/limitations in gpt4 eval. We find OpenChat’s performance on MT-bench is similar to wizardlm-13b. That’s said, there remains a significant gap between open models and GPT-3.5, which is exactly what we aim to emphasize with MT-bench - to highlight this discrepancy. Though not flawless, it’s one step towards a better chatbot evaluation. Please check out our paper/blog for more technical details and leaderboard for complete rankings.
Jim Fan: For most of the “in the wild” trials, GPT-3.5 just feels much better than open-source models that claim good performance metrics. Such “vibe gap” is typically caused by inadequate benchmarking. Don’t get excited by numbers too quickly. Beware of over-claims.
Links: Blog, Leaderboard, Paper.
Falcon-40B is down at 5.17.
Note that reasoning is the place where GPT-4 has the largest edge.
Will they offer all that mundane utility for free? David Chapman thinks that without a moat no one will make any money off of LLMs. Other than Nvidia, of course.
Will Manidis: the core innovation of Foundation Model providers is not technical it’s allowing VCs to deploy $500m into a deal with basically zero underwriting that’s $20m in fees, $100m in carry at a 2x for like . 10 days of memo writing and no customers to reference.
David Chapman: Regardless of how useful GPTs turn out to be, I’m skeptical anyone makes much money off of...