Not sure why that surprises you. These LLM AI models aren’t really intelligent like we think of it.
Basically they’re an outgrowth of the same technology that lets a visual AI model be trained with a whole lot of pictures of clouds, and then be able to identify clouds, or maybe be trained on a bunch of data that shows patterns and be able to fairly accurately predict things. The big difference is that they’re specifically engineered to deal with language data - i.e. take a prompt in, parse it, and produce output in sensible, correct and intelligible language.
What they don’t have is the ability to evaluate which answer that fits the prompt is better or more valid, or whatever. It’ll find an answer that satisfies the prompt (however it does that), and spit the response back in a very coherent fashion. Nor do they generally draw conclusions about data elements- if there’s a relationship between things or a conclusion drawn in the data that it’s trained on, it will return that, but it won’t actually draw those conclusions on its own.
They’re only as good as what you train it with. If you took an LLM and fed it nothing but MAGA and right-wing stuff and asked it political questions, it would spit back coherent, grammatically correct and well formed language, but the content would be that same MAGA stuff you fed it. And if you fed it both sides, it would likely spit back both viewpoints or somehow combine them. It doesn’t have an opinion or any way to judge between the two.
I suppose you could come up with a different sort of AI to actually do that, and then integrate that, but from what I understand, we’re not there yet.
That’s why when you ask it something like “European cities at the same latitude as Portland, Oregon”, it’s going to parse that into whatever it comes up with for european cities, portland, and the same latitude.
And what you got back was entirely sensible in that it’s well constructed, and each part makes sense- cities at the same latitude may not have the same weather, and equatorial areas have more direct sunlight and are hotter than more northerly ones.
But the AI doesn’t really understand the relationship between the two. And really, why would it? It’s not really set up to do that - it’s just spitting back facts about latitudes and weather.