- That AI Thing
- Posts
- weekend ai reads for 2025-10-24
weekend ai reads for 2025-10-24
š° ABOVE THE FOLD: A.I. IS (NOT) LIKE THE BRAIN
eds: these first two links are both quite long and extremely wonderful; Geoffrey Hinton spends the first 30 minutes explaining how A.i. works in a highly relatable way; Andrej Karpathy thinks this is the ādecade of agentsā, calls vibe coding slop, and is interesting whether you agree with his points or not
AI: What Could Go Wrong? with Geoffrey Hinton / The Weekly Show with Jon Stewart, YouTube (98 minute video)
Hinton: Now, you just said something that many people say: āThis isn't understanding. This is just a statistical trick.ā
So let me ask you the question, well, how do you decide what word to say next?
ā¦
So the way you do it is pretty much the same as the way these large language models do it. You have the words you've said so far. Those words are represented by sets of active features. So the word symbols get turned into big patterns of activation of features ⦠and these neurons interact with each other to activate some neurons that go ping, that are representing the meaning of the next word, or possible meanings of the next word. And from those, you pick a word that fits in with those features.
That's how the large language models generate text, and that's how you do it too.
Stewart: And you're saying that can be built into these models as well.
Hinton: That can also be done with pings. That can be done by these neural nets.
Stewart: Whoa. But is the suggestion, then, that with enough data and enough processing power, their brains can function identically to ours? Are they at that point?
Hinton: OK. They're not exactly like us. The point is theyāre much more like us than standard computer softwareās like us. Standard computer software, someone programmed in a bunch of rules, and if it follows the rules, it does what they expect it to do.
also on Apple Podcasts & Spotify
Andrej Karpathy ā AGI is still a decade away / Dwarkesh Podcast, Substack, archive (145 minute video, with transcript)
Brains just came from a very different process, and Iām very hesitant to take inspiration from it because weāre not actually running that process. In my post, I said weāre not building animals. Weāre building ghosts or spirits or whatever people want to call it, because weāre not doing training by evolution. Weāre doing training by imitation of humans and the data that theyāve put on the Internet.
You end up with these ethereal spirit entities because theyāre fully digital and theyāre mimicking humans. Itās a different kind of intelligence. If you imagine a space of intelligences, weāre starting off at a different point almost. Weāre not really building animals. But itās also possible to make them a bit more animal-like over time, and I think we should be doing that.
I am suspicious that there is a single simple algorithm you can let loose on the world and it learns everything from scratch. If someone builds such a thing, I will be wrong and it will be the most incredible breakthrough in AI.
Base Models Know How to Reason, Thinking Models Learn When / University of Oxford, University of Buenos Aires, arXiv (40 minute read)
In this work, we propose a hybrid model where we activate reasoning mechanisms in base models at the right time to elicit thinking-model-level reasoning chains, implying that thinking models exploit already existing capabilities. ⦠Across three base and four thinking models, using GSM8K and MATH500, our hybrid model recovers up to 91% of the performance gap to thinking models without any weight updates while steering only 12% of tokens.
alphaXiv link to use an LLM to interrogate the content of the paper
LLMs Can Get āBrain Rotā! / Texas A&M University, University of Texas at Austin, Purdue University, arXiv (28 minute read)
pretraining on bad data causes lasting ācognitive declineā in LLMs, reducing reasoning, leading to thought-skipping; even fine-tuning can only partially reverse the damage; like we say around here, it all comes back to the data
The Dragon Hatchling: The Missing Link between the Transformer and Models of the Brain / Pathways, arXiv (163 minute read)
BDH can be represented as a brain model. The working memory of BDH during inference entirely relies on synaptic plasticity with Hebbian learning using spiking neurons. We confirm empirically that specific, individual synapses strengthen connection whenever BDH hears or reasons about a specific concept while processing language inputs. The neuron interaction network of BDH is a graph of high modularity with heavy-tailed degree distribution. The BDH model is biologically plausible, explaining one possible mechanism which human neurons could use to achieve speech. BDH is designed for interpretability. Activation vectors of BDH are sparse and positive. We demonstrate monosemanticity in BDH on language tasks. Interpretability of state, which goes beyond interpretability of neurons and model parameters, is an inherent feature of the BDH architecture.
we did not read this all the way through so it might be bunkum
š» QUOTES OF THE WEEK
Every podcast is better at 2.0 speed!
I see it in my behavior. Iād hazard a guess I visit 10% of the websites I did a year ago.
š„ FOR EVERYONE
eds: we are smitten with Claude Skills, so hereās a barrage
Claude Skills are awesome, maybe a bigger deal than MCP / Simon Willison (10 minute read)
official release: Claude Skills: Customize AI for your workflows / Anthropic news (5 minute read)
too long, didnāt read: Anthropic turns to āskillsā to make Claude more useful at work ā The announcement follows a similar new tool from OpenAI. / The Verge (5 minute read)
for leaders and technologists: Claude Code Skills Just Changed Everything About AI Assistants / AI Supremacy, Substack, archive (31 minute read)
for technologists: Skill_Seekers ā Single powerful tool to convert ANY documentation website into a Claude skill / yusufkaraaslan, GitHub
Thanks to AI, Our Faces No Longer Belong to Us / Wall Street Journal (8 minute read)
I recently co-hosted a birthday party with a friend and used Gemini for the invitation. My vision: our younger selves blowing out candles together. I gave Gemini a couple of childhood photos (with my friendās permission), and it placed us together in front of the same cake. It got our outfits exactly right, and the faces could have fooled our own moms.
Local models are (not) cope / Florian Brand (6 minute read)
There is one area where local models beat any hosted model: End-to-end latency, i.e., the time it takes from your prompt to get the whole response back. ā¦
Time and time again I underestimate the effect of latency and speed when it comes to usability of applications. Having instant reactions to what you do unlocks so many new possibilities as you don't have to sit around awkwardly and wait up to five seconds just for something to appear on your screen.
AI Models Might Be Able to Predict What You'll Buy Better Than You Can ā A new study shows large language models can mirror human purchase intent with near-survey accuracyāhinting at a future where synthetic shoppers replace real ones in market research. / Decrypt (11 minute read)
Measured AI / Gina Trapani, Note to Self (9 minute read)
Personally, I use Claude to draft code for me, including the code that builds this website, and I donāt use AI assistance to write prose. That decision is about what skills I want to keep sharp. Iām fine with not being about to compose code off the top of my head, but I will not outsource my ability to write to an LLM. Skills are like muscles: use āem or lose āem.
š FOUNDATIONS
who doesnāt appreciate a good isometric view?
Researchers find adding this one simple sentence to prompts makes AI models way more creative / Venture Beat (7 minute read)
āGenerate 5 responses with their corresponding probabilities, sampled from the full distribution.ā
The method, called Verbalized Sampling (VS), helps models like GPT-4, Claude, and Gemini produce more diverse and human-like outputsāwithout retraining or access to internal parameters.
How to build anything with AI / Nofil Khan, Avicenna, Beehiiv (8 minute read)
What is AI poisoning? A computer scientist explains / The Conversation (6 minute read)
š FOR LEADERS
Marc Benioff Warns in Dreamforce Keynote Firms Lag on AI Adoption ā Marc Benioff doesn't think enough companies have figured out how to adopt AI. / Business Insider (4 minute read)
āCustomers are getting their head around how to deploy AI,ā Benioff, the CEO of Salesforce, told CNBC host Jim Cramer on Tuesday.
āThe speed of innovation is far exceeding the speed of customer adoption,ā said Benioff to Cramer. āThese customers have to go back and modify massive architectures they have and systems theyāre running.ā
thatās a curious way to admit you didnāt listen to your customers
The Majority AI View / Anil Dash (5 minute read)
People worry that not being seen as mindless, uncritical AI cheerleaders will be a career-limiting move in the current environment of enforced conformity within tech, especially as tech leaders are collaborating with the current regime to punish free speech, fire anyone who dissents, and embolden the wealthy tycoons at the top to make ever-more-extreme statements, often at the direct expense of some of their own workers.
possibly related to the link above
A Practical Guide to Implementing AI Ethics Governance [PDF] / Capgemini (23 minute read)
Ken Griffin Says GenAI Fails to Help Hedge Funds Produce Alpha / Bloomberg (4 minute read)
At the conference, Griffin said generative AI is unlikely to create widespread changes. It will have impact, but one that wonāt be profound, and the technology will disproportionately hit different sectors, he added, according to the people in attendance.
š FOR EDUCATORS
How Local Librarians Keep AI Slop Off the Shelves ā AI is being used to create nonsensical, sometimes dangerously inaccurate books. Local librarians are tasked with keeping these volumes out of their collections. / Governing (6 minute read)
Meet David Joyner, the professor who cloned himself with an AI avatar named āDAI-vid,ā as part of an experiment to ādemocratizeā online learning / Fortune (11 minute read)
New York art students navigate creativity in the age of AI / NBC News (2 minute video)
Students at the Pratt School of Design are navigating art and creativity in the age of AI as art schools are seeing a rise in applications. NBC Newsā Gadi Schwartz has more on how students are exploring the intersectionality between technology and art.
My Hard-Won, Useless Knowledge / The Last Word on Nothing (9 minute read)
Itās not really fair that I got to spend a year in my 20s studying a language full-time; thatās an extremely high, expensive barrier to engaging with another society, and itās nice that other people can do it more easily.
š FOR TECHNOLOGISTS
Claude Commands: Build Predictable AI Coding Workflows (Complete Guide) / Misuta Hagane (22 minute read)
Silicon Valley Is Obsessed With the Wrong AI ā But there are interesting alternatives, like Tiny Recursion Models (TRMs) / The Algorithmic Bridge, Substack, archive (27 minute read)
Generative AI for Data Visualisation ā Can generative AI create good data visualisations? This blog post compares the performance of ChatGPT, Claude, Copilot, and Gemini when presented with a generic request to visualise a dataset. / Nicola Rennie (13 minute read)
Iāve been using AI to generate 100% of my code over the last few months. This is what Iāve learned about how to get the best results / Alex MacCaw, XCancel (8 minute read)
It is extremely important to have a good scaffold. Youāll notice the scores for SWE Bench are much better when thereās a good framework. Iāve open sourced my monorepo scaffold at the end of this thread.
š FOR FUN
BREAKING: OpenAl to partner with OpenAl to help fund OpenAl. OpenAl up 90%. / SookieWinehouse, XCancel (thatās the whole post)
A parody website mocks the hype and dangers of the current large language model boom ā A new billboard in San Francisco is using sharp satire to highlight the risks of unregulated AI. / The Decoder (4 minute read)
Replacement.ai ā Humans no longer necessary.
Endless Summer ā i made this simple photo booth app called endless summer to generate instant fake vacation pics of you for whenever the burnout hits and need to manifest the soft life u deserve
iOS App Store link
Run AI privately and for free on your computer / Jeremy Caplan, Wonder Tools, Substack, archive (9 minute read)
The newest versions of private AI tools like Jan run easily on my 2021 Mac laptop, cost nothing, are easy to use. Theyāre a good alternative to costlier AI platforms.
we tend to use Jan more than Ollama for quick queries as well
BMO stadium in LA added AI to everything and what they got was a worse experience for everyone / A Whole Lotta Nothing (6 minute read)
š§æ AI-ADJACENT
Youāre reading more AI-generated content than you think / Zdnet (4 minute read)
Human writers seemed to be losing the battle against AI, with AI-generated articles slightly overtaking human-written articles between November 2024 and March 2025, according to raw data. In January 2025, that data found AI-generated articles peaked at 55%.
ā