Discussion about this post

User's avatar
Moral Particle's avatar

I stumbled across your Substack looking for thoughtful reactions to Leopold Aschenbrenner’s Situational Awareness document, and now I’m reading older posts. I apologize for this somewhat untimely comment. (I say “somewhat” because we’re still waiting for GPT-5!) In general, I agree with much of what you say here and how you’re framing the issues.

One thought on current models: Outside of the world of education, it’s hard to appreciate how significantly and quickly LLM chatbots, as flawed and limited as they currently are, have changed academic writing and grading. Take-home essays and papers, which used to be assigned by the tens of millions from middle school through graduate school, have been essentially destroyed as a means of evaluating student progress and achievement. Teachers, professors, and administrators are still working through this paradigm shift, but the world has forever changed. Less acknowledged, but just as real, will be the effects on “the publish or perish” regime of academia, especially in the Humanities and “social sciences.” As less tech-savvy but stressed and highly-motivated graduate students and junior faculty realize that they can now load published journal articles and even entire books into context windows and get almost instantaneous draft articles from LLMs analyzing those texts in the style of academic writing with only a few prompts, the floodgates will open. The economic impact of that might be hard to quantify, but it’s a profound shift in the (giant!) Educational-Industrial Complex.

A bigger and more general issue is hallucination. I haven’t been able to determine whether and how future models will avoid hallucinations. Although well-documented and discussed as to current models, the discussions I’ve seen about AIs of the future don’t address the issue. It’s a huge problem! Right now, even the best LLMs will make up - literally invent - facts, papers, biographies, even people. I’m most familiar with how lawyers using LLMs get briefs that cite fake cases and statutes. I can’t overemphasize how unacceptable it is to submit an argument with false, made-up authority to a judge or panel of judges! Similarly, a colleague in academia has given me numerous examples of LLMs referring to fake journal articles and books. According to him, current models are “worse than useless” for research assistance because the output is so untrustworthy. Will more compute and more training data eliminate the problem? I know that AI researchers are well-aware of the problem, but I hear surprisingly little about potential solutions even as they spin grand visions of “drop-in knowledge workers” performing human-level work. A drop-in knowledge worker who provides fake legal authority (or fake medical studies) is dangerous, not useful, even if the error rate is low. (In many cases now, the error rate is high!) To be clear, I imagine - indeed, think it’s likely - there are solutions, but I’m surprised at how under-discussed this issue is. If you have knowledge or thoughts about this, it would be great to hear them!

Expand full comment
Tony Asdourian's avatar

Great blog title! I sort of expected what you were going to say in this post, as it mirrors what I have noticed following AI for the last year— lots of flashy announcements and noise, but nothing that seems to change things RIGHT NOW. I am surprised, though, that you think it is 5-40 years for Agentic AI— more than anything else, that prediction makes me think you perceive the current approach as fundamentally limited, even if they can milk out some superficially impressive improvements for GPT-5 and 6.

Expand full comment
6 more comments...

No posts