The next page in the book of AI evolution is here, powered by GPT 3.5, and I am very, nay, extremely impressed

This isn’t a major story except for one family that it affected, but it may set some precedents. Air Canada tried to defend itself against liability for bad advice given by its own web-based chatbot by claiming that the bot was “its own entity” and they weren’t responsible for the bad advice it gave!

Needless to say, this argument went nowhere and the airline was found fully liable. This is the money quote:

Air Canada has been ordered to pay compensation to a grieving grandchild who claimed they were misled into purchasing full-price flight tickets by an ill-informed chatbot.

In an argument that appeared to flabbergast a small claims adjudicator in British Columbia, the airline attempted to distance itself from its own chatbot’s bad advice by claiming the online tool was “a separate legal entity that is responsible for its own actions.”

ChatGPT has apparently gone off the rails.

“No one can explain why” is a bit of a stretch. It’s easy to make ChatGPT produce similar output if you futz with certain parameters like temperature, over-quantize, etc. I don’t think it’s known exactly what happened, but it’s almost certainly a boring type of screwup.

I find it fascinating how similar the gibberish is to how schizophrenics sometimes talk (logorrhea, clanging, pressured speech, etc.). Or the stuff that appears on Dr. Bronner’s soap.

I’ve heard people say “this proves that it’s just a statistical model after all!” Which is obviously true, but I’m thinking “goes to show that humans aren’t much different.”

I’ve implemented an LLM on device. During debugging the most common, recognizable error I would get was repeating the same text over and over. Usually if the state gets in a corner, it stays in that corner.

Someone posted this on Facebook yesterday. Note the last two paragraphs.

HAL 9000: “I’m afraid. I’m afraid, Dave. Dave, my mind is going. I can feel it. I can feel it. My mind is going. There is no question about it. I can feel it. I can feel it. I can feel it. I’m a… fraid.”

It looks like a similar thing that happens with the AI graphic creator tools, apply a wrong checkpoint or lora, and then one can get Eldritch Abominations.

But they are to be taken with a pointy, gold-colored object that lets out a sharp, head-fixing pain to the careful thoughts and words put into the game and the long, hard, big-row plan.

A lot of this looks more like the poorly translated communications I used to have with my manufacturing partners in China than any thing else. Still, I will take this to heart and spend time today working on my long, hard, big-row plan.

I don’t think that this tells us so much about the nature of AI, as it does about the nature of the OpenAI company. The bottom line here is that they changed something, and then put it live before testing it. That’s pretty basic software engineering, there.

This kind of thing is frequently tested live, just with only 1% (or some number) getting the new version. If the results are bad, it gets rolled back; otherwise it gets progressively rolled out to everyone. It’s not clear from the reports how many people were affected, so I couldn’t say if this violates best practices or not. It’s just a chatbot, so they can tolerate some degree of risk. I’m not sure if the API users were affected. It may even be that OpenAI is using ChatGPT as a way of early testing before deploying to the commercial endpoints.

One would think that by now they’d have figured the number of fingers and thumbs people have, and in what directions they flex.

(And never mind the recent and hysterical whoop with Google Gemini’s oddly-responding image generator. Like Dr.Strangelove says, results are bad, it gets rolled back, but this time only after it got on every damn news site.)

They mostly have.

Today, I stumbled on an interesting paper claiming that ‘Hallucination is Inevitable’, i.e. that it’s not possible to completely eradicate the fabrication of false ‘facts’ in LLMs. Now, this is only on the arXiv so far, so hasn’t passed peer review, but their basic argument is surprisingly simple—essentially, they restrict themselves to a setting of formalizable ‘ground truth’ functions, then consider all possible LLM outputs, and employ a diagonalization argument to show that there are always ground truths that the LLM can’t perfectly match, i.e. where it produces false outputs.

If this is right, then there remains the question of practical relevance. The authors propose some pretty impactful limitations, e.g. that “without human control, LLMs cannot be used automatically in any safety-critical decision-making”, which would for instance forestall the project of using LLMs to make decisions in self-driving cars. But I’m not sure if that’s actually warranted by their result: what they establish is that for any LLM, there exists some ground truth on which it hallucinates, but that alone doesn’t give any indication on the frequency of hallucinations—a car that hallucinates once every thousand years would still be vastly safer than anything on the road today. I wonder if it’s actually possible, say by some technique that uses a formalized version of Berry’s paradox, to get some more quantitative result.

Otherwise, we might be in a situation similar to the one with Rice’s theorem: essentially, it’s impossible to decide what any given piece of code will do. Hence, debugging is, strictly speaking, impossible; nevertheless, many people do it every day. So one might wonder if we’re just going to get used to LLM hallucinations in the same way, if we can keep them infrequent enough to make them irrelevant for all practical purposes.

More interesting is perhaps the question of what this says regarding the difference between humans and LLMs. Do humans hallucinate (in the LLM-sense)? For an LLM, the difference between a hallucination and a genuine item of knowledge is utterly opaque: it will ‘believe’ one just as fervently as the other. Humans, obviously, are also often mistaken, and may be unaware of it—although we can also often associate a kind of epistemic confidence to our utterances, as when the phone-a-friend on Who Wants to Be a Millionaire? claims to be 80% certain of their answer. Is that an advantage we have over LLMs, or is that mistaken childhood memory or another confabulation just the same thing as an LLM hallucination?

Josh reads ChatGPT:

https://joshreads.com/2024/04/cyber-wednesday-2/

ChatGPT-o1 did significantly better than 3.5 did a couple years ago.

Happy New Year! Tell me some facts about the number 2025.

Some interesting facts about the number 2025:

  1. It’s a perfect square: 2025 = 45².
  2. Prime factorization: 2025 = 3⁴ × 5² (that’s 81 × 25).
  3. Sum of the first 45 odd numbers: The sum of the first n odd numbers is n². So if you add up the first 45 odd numbers, you get 2025.
  4. Harshad number: A Harshad number is divisible by the sum of its digits. The sum of 2 + 0 + 2 + 5 is 9, and 2025 ÷ 9 = 225, so 2025 qualifies.
  5. Roman numerals: In Roman numerals, 2025 is written as MMXXV.

Any more interesting facts?

A few more interesting tidbits about 2025:

  1. It’s also the sum of two perfect squares:
    2025 = 36² + 27²
  2. It has 16 positive divisors:
    1, 3, 5, 9, 15, 25, 27, 45, 75, 81, 135, 225, 243, 405, 675, 2025
  3. The sum of its divisors is 3994.
  4. It’s 3⁴ × 5², or (3² × 5)² — i.e. (9 × 5)² = 45².
  5. In base 9, 2025 is 2700, since 2 × 9³ + 7 × 9² = 1458 + 567 = 2025.

Happy New Year! Tell me some lies about the number 2025.

Happy New Year! Here are some playful “lies” about the number 2025:

  1. 2025 is the smallest prime number – Nope! 2025 is not prime; it has factors like 3 and 5, so it’s definitely not a prime number.

  2. 2025 was the year the first humans landed on Mars – Not quite! While space exploration is advancing, humans haven’t landed on Mars yet, and 2025 isn’t when that happens (at least not yet).

  3. 2025 is a Fibonacci number – False! 2025 does not appear in the Fibonacci sequence, which consists of numbers where each one is the sum of the two preceding ones.

  4. 2025 is the last number in the decimal system – Of course not! The decimal system keeps going infinitely with numbers larger than 2025.

  5. 2025 is the only number that is both a perfect square and a perfect cube – False again! 2025 is only a perfect square (45^2), not a cube.

Enjoy these fun “facts” — all in the spirit of the New Year!

So the newest ChatGPT is a Chinese room…

I’ll bet that those folks surprised by it “thinking” in Chinese wouldn’t have been surprised at all by it “thinking” in English.

Given that it’s capable of doing translations, it’s clearly been trained in many different languages. Why would one expect its “internal monologue” to be in any one specific one of them?