From the article:
AI chatbots have become an integral part of the way information is disseminated, processed, and consumed. However, their growing presence is not without significant challenges. A key issue is the surprising level of inaccuracy in the information they provide, as highlighted by a study from the Tow Center for Digital Journalism. This study revealed that AI chatbots, including prominent ones like ChatGPT and Gemini, offer wrong answers over 60% of the time when tasked with sourcing news excerpts. This tendency towards inaccuracy isn’t just a technical glitch but raises serious concerns about the reliability of AI systems in critical applications like news reporting, where accuracy is paramount.
The consequences of these inaccuracies are far-reaching. They not only spread misinformation but also damage the trust between the public and the media, as well as between publishers and the companies developing these AI tools. For instance, chatbots have been known to fabricate headlines and fail to attribute articles correctly, often linking to unauthorized or incorrect sources. This not only misleads the public but also harms news publishers’ reputations and revenue potential by diverting traffic and diminishing the perceived value of credible news outlets.
…
AI search engines, powered by sophisticated algorithms, are increasingly shaping the way people access and consume information. However, a critical analysis reveals that these AI-driven platforms often suffer from significant inaccuracies, especially when sourcing news content. According to a recent study by the Tow Center for Digital Journalism, AI chatbots like ChatGPT and Gemini are frequently incorrect, providing inaccurate information over 60% of the time when tasked with identifying correct news sources. This high error rate stems from tendencies to fabricate headlines, incorrectly attribute article sources, and link to erroneous or unauthorized sources, thereby affecting the integrity and reliability that users expect from such advanced technologies. This issue not only misleads users but also poses substantial threats to the credibility and operational viability of news publishers. For further details on these concerns, you can refer to the study findings .
The repercussions of AI search engine inaccuracies extend beyond professional journalism to encompass broader information ecosystems. When AI models consistently disseminate false or misleading information, they contribute to the erosion of public trust in both AI tools and traditional media outlets. This erosion is not just theoretical but has tangible consequences: misinformation can shape public discourse, influence political perceptions, and even sway election outcomes if unchecked. As these technologies continue to evolve and integrate more deeply into daily life, the imperative for transparency, accuracy, and accountability in their operation grows ever more critical. A detailed examination of these issues and their implications for news dissemination can be explored through the comprehensive research by the Tow Center for Digital Journalism on AI inaccuracies .
Of course, these isn’t aren’t just limited to sourcing and referencing news articles (something even a marginally mature ‘AI’ should be capable of doing) but to accessing, summarizing, and providing important information from any source.
Quoting a bullshit generator to argue that it is not quite as bullshitty as clear evidence would make it seem is about as sensible as fucking for virginity. You keep arguing that chatbots are an “amazing and almost immeasurably important and useful technology” and that it is just a “people problem” that “tendency for naive users to believe them to be dependable”, but in fact in order to be used as a general knowledge system they have to be dependable, at least to the extent that an actual expert is, and they absolutely need to not fabricate sources or citations so that a user can at least know where to look to verify that the source of information is correct and make some effort to verify that the interpretation is what the chatbot says it to be.
The reality is that the use case is for non-experts doing some kind of information research, many of whom (like the poster who used a chatbot to ‘solve’ a physics problem discussed upthread) don’t have enough information to verify a correct answer or quickly intuit a wrong answer, and so they will generally take whatever answer they are given as essentially correct, possibly posting it into a work product, paper, or news article where it will be disseminated to the wider world. That LLM-based chatbots can’t even be used to take food and beverage orders off of a limited menu indicates that the technology is nowhere near mature enough to be deployed to the public and certainly should be used in any way to produce safety-, security-, or health-critical information or directions.
Stranger