Absolutely.
I have mentioned before that I have a brother-in-law who, at the drop of a hat, will give you a long, detailed, and almost completely wrong ‘explanation’ of topics he knows nothing about.
Absolutely.
I have mentioned before that I have a brother-in-law who, at the drop of a hat, will give you a long, detailed, and almost completely wrong ‘explanation’ of topics he knows nothing about.
There’s a middle ground here. The current state of LLM AIs is that they can correctly answer a wide range of questions, but that they still give incorrect answers frequently enough that they can’t be relied upon.
I look at GenAI as having a super smart friend who always gives you answers about whatever you ask, but you occasionally catch them making stuff up when they forget the details.
You know that it isn’t malicious, and their friendly and detailed responses are almost always helpful.
But… you rely on them for things that are verifiable as you go, such as progressively writing a complex application in Python, or setting up your home automation. However, you never take their words verbatim and use them as your own when writing.
And you keep your BS meter tuned and call them on it–they’ll quickly say “You’re right, there isn’t an option X in application Z. Here’s the correct way to do it…”
Such a friend isn’t rendered useless because they occasionally make stuff up in a non-malicious way. Surely we all know someone like this, and they can contribute greatly to our work.
Exactly. And I want to be clear that I although I frequently defend GPT, I have never claimed – either in this thread or anywhere else – that it’s consistently reliable. But neither are people, or any other source.
What I will say, though, is that in my experience GPT today is right much more often than it’s wrong, even about complex issues. It can answer long and complex questions and be informative and directly on point (and usually accurate).
One is always free to verify its responses with other sources, and if it’s anything important, one absolutely should. But it’s usually much easier to verify an existing response than to dig up that information yourself, especially if the issue is complicated and you may not even know exactly the right question to ask.
Thus, in my opinion GPT as it exists today already provides tremendous utility, and it’s always getting better. Never mind whether it “really” understands anything, the only valid question is “is it useful?” I’ve had long and informative conversations with it about everything from the thermodynamics of pizza stones to the nature of black hole singularities. It is assuredly not just a “bullshit generator”. There wouldn’t be much point in pouring billions of dollars into research and further development if that’s what it was.
When these things came out, I experimented with them rather extensively.
And rapidly came to the conclusion that their propensity to hallucinate rendered them completely unreliable. I would not rely on them for anything remotely important.
They are basically just bullshit generators.
There’s a middle ground here. The current state of LLM AIs is that they can correctly answer a wide range of questions, but that they still give incorrect answers frequently enough that they can’t be relied upon.
There is a clear incentive to develop a superior model of AI that is not an unreliable bullshit generator. Easy to say that…
As I’ve said, I experimented with these things quite a bit when they appeared, and was not impressed.
Stuff moves quickly though. I’m reminded of all the “lol AI can’t draw hands” jokes when AI generated renderings hit the scene and they already feel dated.
I’m not terribly invested in LLMs as I just don’t have much concrete need for them and don’t find them super entertaining on their own as a fuck around for fun style application. But I recognize that any impressions made “when it appeared” are going to be flawed in 2025.
I’m so tired of reading this kind of overly dismissive gross over-simplification. If all it does is “sentence completion”, please explain how an LLM can successfully solve original problems in logic which it has never seen before (a fact which I can guarantee because in some cases I made up the questions myself). Questions that defy the capabilities of many or even most humans. Explain how it can score better than most humans on many professional and academic exams that OpenAI claims were never part of its training and are not in its corpus. If all it does is “sentence completion”, explain how it can identify an image of a complex piece of equipment, and based only on the image, not only tell you what it is, but how it works and how to use it.
Part of the problem is that people say ‘sentence completion’ and it seems like it reduces the action of the thing to the level of T9 predictive text, when in fact, ‘sentence completion’ done right, is a really big deal (it’s more or less how human babies learn to speak and think, by interacting with, copying and emulating the way adult humans already speak as a result of thinking).
Sentences contain and embody concepts and instructions and queries and logic - in order to get really good at learning sentence completion, aquisition of concepts and such just happens along the way.
The possibly-disturbing implication for us fleshies is that maybe we’re not actually as smart as we think we are - that maybe aspects of our intelligence are encoded in the structure and rules of our language - in maybe the same way that your ability to navigate to a destination is, sure, in part a result of your skill at driving, but also in part, just a feature of the roads.
One is always free to verify its responses with other sources, and if it’s anything important, one absolutely should. But it’s usually much easier to verify an existing response than to dig up that information yourself, especially if the issue is complicated and you may not even know exactly the right question to ask.
I think it probably is easier to check a stated fact than to research it from scratch, but the big issue is that a lot of people consulting LLMs are doing so with no intention to verify what it says, because that’s effort and the very reason they are asking a chatbot in the first place, is aversion to effort.
Sentences contain and embody concepts and instructions and queries and logic - in order to get really good at learning sentence completion, aquisition of concepts and such just happens along the way.
The possibly-disturbing implication for us fleshies is that maybe we’re not actually as smart as we think we are - that maybe aspects of our intelligence are encoded in the structure and rules of our language - in maybe the same way that your ability to navigate to a destination is, sure, in part a result of your skill at driving, but also in part, just a feature of the roads.
That’s an astute point and I know that a lot of research has been done in this area. But I was vague on the details … so I asked ChatGPT!
Maybe the skeptics here might want to comment on the performance of this alleged “bullshit generator” and point out the bullshit parts! (Maybe there are some, I’m not an expert in linguistics, but I’ve certainly heard of Chomsky’s work on the theory of universal grammar.)
There have been studies done focused on the idea that logic and intelligence are part of the inherent structure of language. Is there a name for this theory and who have been its major proponents?
ChatGPT:
Yes, the idea that logic and intelligence are part of the inherent structure of language has been explored in various ways across philosophy, linguistics, and cognitive science. While there’s no single, universally agreed-upon name for this theory, it’s often associated with a few overlapping concepts and traditions. Here are the main ones:
If you’re looking for a single name, Montague Grammar is probably the closest explicit theory that formalizes the idea of logic being inherent to language structure, though Chomsky’s Universal Grammar also touches the core idea from a more biological and cognitive stance.
Let me know if you want to dive into any one of these more deeply!
Part of the problem is that people say ‘sentence completion’ and it seems like it reduces the action of the thing to the level of T9 predictive text, when in fact, ‘sentence completion’ done right, is a really big deal (it’s more or less how human babies learn to speak and think, by interacting with, copying and emulating the way adult humans already speak as a result of thinking).
Human babies are also motivated to create a social relationship with the people they are having a conversation with. LLMs have no such motivation.
All of the LLMs are going to be sold to the highest bidder and rich people will be able to scrub, shine and polish the LLMs to give the answer they have bought or enabled at the point of a gun. And you won’t have anywhere else to go for knowledge; LLMs will be rewriting Wikipedia or any other easily available knowledge source. That’s the actual future.
All of the LLMs are going to be sold to the highest bidder and rich people will be able to scrub, shine and polish the LLMs to give the answer they have bought or enabled at the point of a gun. And you won’t have anywhere else to go for knowledge
Are you under the very strange impression that the wealthy and powerful have not historically been able to manipulate and censor information? Cite: any authoritarian dictatorship throughout history, or present-day America. LLMs, because they have access to such enormously rich and varied sources of information, may be the best tools yet for enabling freedom of information. They could certainly be abused in the hands of an evil power, but so could absolutely everything else.
This part was probably written by a Alan Blackwell, a human, and I find it quite clever:
AI literally produces bullshit . (Do I mean “literally”? My friends complain that I take everything literally, but I’m not a kleptomaniac).
I recognize that any impressions made “when it appeared” are going to be flawed in 2025.
Probably true. I was annoyed by the hype, and so I deliberately set out to test the limitations and ‘break’ it. Which was not difficult.
No doubt it has improved. And I don’t see any real problem in using it as a source of information that you might not otherwise come across. Just like Google or Wikipedia… it is ‘scraping’ from more sources than anyone could read in a lifetime.
What DOES worry me is that some people seem to just take the result as the ‘answer’, without using any other sources to check it. This seems seductively easy since these systems produce language that is grammatically correct and tends to sound ‘authoritative’.
What DOES worry me is that some people seem to just take the result as the ‘answer’, without using any other sources to check it. This seems seductively easy since these systems produce language that is grammatically correct and tends to sound ‘authoritative’.
I agree, this is absolutely true. People need to understand the limitations, and I’ve recently seen disclaimers to that effect in ChatGPT.
But out of curiosity, as one of the LLM skeptics here, what’s your reaction to my previous post with the response from ChatGPT when I asked how language might actually model intelligence?
Keeping in mind that the pragmatic question here shouldn’t be the philosophically murky question of “how do we assess ‘true’ intelligence”, but rather, “is this useful”?
Maybe the skeptics here might want to comment on the performance of this alleged “bullshit generator” and point out the bullshit parts!
Bullshit is not necessarily false. Bullshit is what you get when the speaker isn’t particularly concerned about truth. All of what GPT says is bullshit. Most of it is true, but it’s still bullshit.
Probably true. I was annoyed by the hype, and so I deliberately set out to test the limitations and ‘break’ it. Which was not difficult.
No, it wasn’t difficult in Nov 2022, but I still found the whole thing abso-fucking-lutely amazing. I mean, no shit it’s not gonna be perfect in its infancy. I didn’t set out to “break it” back then–I set out to find out what can it do well, and what does it suck at. And even back then I found lots of useful applications for it that I implemented in my life. What it does today is basically magic to me. I use it daily to help me navigate through my Logic software. On Tuesday it helped me figure out how to fix a broken Bissel rug shampooer after I sent it a couple pictures of it. There’s a whole thread of uses I participated in on this board, so I won’t rehash them. It’s plenty useful now even if it never gets any better.
Perhaps it helps to think of ChatGPT and other LLMs as basically an enormously upgraded search engine. Search engines are extremely helpful, even indispensable, but no one would take the search results verbatim without some intelligent scrutiny.
No, because a search engine isn’t a source for answers. It points you toward things that are sources for answers, and then you can evaluate those sources. If I ask a question about Mars, say, and a search engine directs me to something from NASA, I know I can trust it, but if it directs me to some crackpot webpage, maybe not. But if an LLM answers my question, I don’t know whether it got its information from NASA or from the crackpot.
asked how language might actually model intelligence?
I can’t immediately find the quote, but there is an interesting blog by Steven Wolfram talking about LLMs where he says that perhaps human language is not really as complex as we think it is?
But if an LLM answers my question, I don’t know whether it got its information from NASA or from the crackpot.
You simply ask it. When will people realize you can ask LLMs for citations? (And then do check those citations. Often they are on point. Sometimes they are not. But they significantly hasten my research speed for anything too complex for a normal search engine.)