I don’t know exactly how the ChatGPT crawler works, but I think Discourse is also a special challenge in that its post numbering URL system is kinda “fake”; it is not a normal URL, but a dynamically updated browser history state that’s changed via Javascript as you scroll down the page.
I think probably many crawlers would struggle with it, with or without AI.
That said, though, I think it was able to read the entire thread just fine before you started specifying particular post numbers. My browser also kinda bugged out with the post numberings… post 48 completely disappeared for me (it would jump from 47 to 49) until I reloaded.
Granted, Discourse is common enough that OpenAI probably should’ve written a special parser for it, either manually or with AI help. But still, I wouldn’t necessarily use this particular incident as a representative test of ChatGPT’s overall ability or usefulness. This is just one of the things it happens to be bad at, but it’s still good for other things.
No, I think 4.5 or o3-mini-high are newer. Newer isn’t necessarily better, though. In ChatGPT and most other clients (not sure about Poe), you can usually choose which model you want to use (and switch between them after every every prompt).
OpenAI’s own suggestion (and I concur) is that 4o is better for most queries. o3-mini-high is much better for coding but not as good for anything else. I’m not sure what o3-mini is good for, if anything (maybe speed?). There are also other models I don’t bother with.
The main benefit of the ChatGPT app is that it’s not charged by usage, just a flat $20 a month or so, and it usually gets new models before the API does (meaning before third party clients would). There is also a $200/mo Pro version that’s supposed to be quite a bit better, and includes more features and models, but I’m not rich enough to try it.
Outside of OpenAI, I think the latest Gemini (Google’s AI) is leading many of the benchmarks. Some programmers prefer Claude for coding. Copilot (for Github, not Microsoft’s confusing reused branding of it in Windows) is another option, but seems to have fallen behind. Deepseek is a free Chinese model that rivals some of the best Western ones, but there is some concern over both its censorship/propaganda potential and its development process (suspicion over whether it uses stolen code or weights from other models).
Those are just some of the big names. There are dozens more, including ones you can train or fine tune yourself. We’re still at the height of the hype cycle, so the choices are pretty overwhelming these days, but the analysis, testing, and tutorials have gotten a lot better since the early days.
Personally I just use ChatGPT, more out of habit than anything. The leaders keep changing every month, but I don’t care enough to try to keep up.
As far as I know, yes. There is concern and active research about how to prevent future models from being trained on the poorer regurgitations of prior models. There is fear that the first movers and big companies (OpenAI and Google, in particular) have a huge advantage there by having access to large pre-AI datasets that newer companies cannot easily get anymore. It’s the internet equivalent of low-background steel, I guess.
I can’t say much more than that. AI isn’t my specialization and I am not smart enough to really describe or analyze it much past a layman level.
Maybe subscribe for a month and see if you find any good uses for it? Many people won’t. Some do. Worth trying, since it’s easy to cancel if you don’t end up finding it helpful.