Senior Lecturer in Learning Analytics, University of South Australia Vitomir Kovanovic does not work for, consult, own shares in or receive funding from any company or organization that would benefit from this article, and has disclosed no relevant affiliations beyond their academic appointment. University of South Australia provides funding as a member of The Conversation AU. View all partners Less than two years ago, the launch of ChatGPT started a generative AI frenzy. Some said the technology would trigger a fourth industrial revolution, completely reshaping the world as we know it. In March 2023, Goldman Sachs predicted 300 million jobs would be lost or degraded due to AI. A huge shift seemed to be underway. Eighteen months later, generative AI is not transforming business. Many projects using the technology are being cancelled, such as an attempt by McDonald’s to automate drive-through ordering which went viral on TikTok after producing comical failures. Government efforts to make systems to summarise public submissions and calculate welfare entitlements have met the same fate. So what happened? Like many new technologies, generative AI has been following a path known as the Gartner hype cycle, first described by American tech research firm Gartner. This widely used model describes a recurring process in which the initial success of a technology leads to inflated public expectations that eventually fail to be realised. After the early “peak of inflated expectations” comes a “trough of disillusionment”, followed by a “slope of enlightenment” which eventually reaches a “plateau of productivity”. A Gartner report published in June listed most generative AI technologies as either at the peak of inflated expectations or still going upward. The report argued most of these technologies are two to five years away from becoming fully productive. Many compelling prototypes of generative AI products have been developed, but adopting them in practice has been less successful. A study published last week by American think tank RAND showed 80% of AI projects fail, more than double the rate for non-AI projects. The RAND report lists many difficulties with generative AI, ranging from high investment requirements in data and AI infrastructure to a lack of needed human talent. However, the unusual nature of GenAI’s limitations represents a critical challenge. For example, generative AI systems can solve some highly complex university admission tests yet fail very simple tasks. This makes it very hard to judge the potential of these technologies, which leads to false confidence. After all, if it can solve complex differential equations or write an essay, it should be able to take simple drive-through orders, right? A recent study showed that the abilities of large language models such as GPT-4 do not always match what people expect of them. In particular, more capable models severely underperformed in high-stakes cases where incorrect responses could be catastrophic. These results suggest these models can induce false confidence in their users. Because they fluently answer questions, humans can reach overoptimistic conclusions about their capabilities and deploy the models in situations they are not suited for. Experience from successful projects shows it is tough to make a generative model follow instructions. For example, Khan Academy’s Khanmigo tutoring system often revealed the correct answers to questions despite being instructed not to.
This article was autogenerated from a news feed from CDO TIMES selected high quality news and research sources. There was no editorial review conducted beyond that by CDO TIMES staff.
Need help with any of the topics in our articles? Schedule your free CDO TIMES Tech Navigator call today to stay ahead of the curve and gain insider advantages to propel your business!