Please Note: This email did not come from ANU, Be careful of any request to buy 
gift cards or other items for senders outside of ANU. Learn why this is 
important.
https://www.scamwatch.gov.au/types-of-scams/email-scams#toc-warning-signs-it-might-be-a-scam
`

Payoff from AI projects is 'dismal', biz leaders complain

No wonder most orgs are slowing their spending

By Thomas Claburn  Wed 12 Jun 2024 
https://www.theregister.com/2024/06/12/survey_ai_projects/


Businesses have become more cautious about investing in artificial intelligence 
tools due to concerns about cost, data security, and safety, according to a 
study conducted by Lucidworks, a provider of e-commerce search and customer 
service applications.

"The honeymoon phase of generative AI is over," the company said in its 2024 
Generative AI Global Benchmark Study, released on Tuesday.

"While leaders remain enthusiastic about its potential to transform businesses, 
the initial euphoria has given way to a more measured approach."

Between April and May 2024, Lucidworks conducted a survey of business leaders 
involved in AI adoption in North America, EMEA, and the APAC region. The 
respondents, it's claimed, were drawn from 1,000 companies with 10more 
employees across 14 industries, all of which are said to have active AI 
initiatives underway.

About 23 percent are executives and about 50 percent are managers, with 86 
percent involved in technology decision making. Thirty-nine percent of 
participants hailed from North America, with 36 percent from EMEA, and 24 
percent from the APAC region.


According to the results of the survey, 63 percent of global companies plan to 
increase spending on AI in the next twelve months, compared to 93 percent in 
2023 when Lucidworks conducted its first investigation.

Broken down by region, the highest percentage of those planning to spend more 
was documented among US respondents, at 69 percent. But in the APAC region, 
less than half (49 percent) of Chinese business leaders expect to increase AI 
spending, which is down from last year when every one polled said they would be 
spending more.

Among all organizations, 36 percent planned to keep AI spending flat in 2024, 
compared to just 6 percent last year.

The slowdown can be attributed to several factors.

One issue is that AI has not yet paid off for those trying to make it work. 
"Unfortunately, the financial benefits of implemented projects have been 
dismal," the study says. "Forty-two percent of companies have yet to see a 
significant benefit from their generative AI initiatives."


And to realize meaningful benefits, it's necessary to move past the pilot 
testing stage, something few companies have managed. Only 25 percent of planned 
generative AI investments have been completed to date, the study says.

Another reason is growing concern about the cost of AI projects, up 14x since 
last year. There are also many more worries (5x more) about the accuracy of 
responses provided by AI systems.

The costs involved in using generative AI strategically can really add up

With regard to cost, some 49 percent of organizations have opted to use 
commercial LLMs like Google's Gemini and OpenAI's ChatGPT. Another 30 percent 
use both commercial and open-source LLMs, while just 21 percent have bet 
exclusively on open-source LLMs like Llama 3 and Mistral. The Lucidworks study 
predicts the balance will shift toward open-source LLMs based on anticipated 
open-source model performance gains and on cost considerations.

"Absolutely the costs involved in using generative AI strategically can really 
add up, regardless of whether you're hosting your own large language models or 
using commercial APIs," said Eric Redman, senior director of product for data 
science at Lucidworks, in an email to The Register.

"These initial costs are comparable in the grand scheme of things, but they're 
really just the tip of the iceberg."

According to Redman, security and accuracy, response alignment with policies, 
data acquisition costs, and keeping costs under control all need to be 
considered.

"The bottom line is that ensuring AI security, accurate AI responses, and 
responsible data acquisition all come with a price tag," said Redman. "Cutting 
corners in these areas can lead to inaccurate or inappropriate responses, which 
ultimately undermines the value and effectiveness of your AI implementation."

Among the organizations surveyed, the best generative AI initiatives involved 
governance (standardizing models to ensure alignment, limiting access to 
generative AI tools based on role, and so on) and cost reduction, both general 
and administrative (Q&A testing, debugging, code suggestion, and HR help 
documentation).

The survey indicates that qualitative applications, which use text and provide 
a narrow response, have been the most successful generative AI initiatives, 
accounting for about a quarter of successful implementations. Specifically, 
these include projects for generating FAQs and providing HR support. And they 
tend to be the simplest to implement.

Applications with a quantitative component – using generative AI to monitor, 
predict, analyze, optimize, prioritize, and other more challenging projects – 
have had a rougher time, with less than 15 percent successfully implemented. As 
the study observes, these involve optimizing search results, screening job 
applicants, supporting financial result closure, and so on.

Redman said code generation's popularity as a top use case makes a lot of 
sense. "It's essentially a prime example of how AI copilots can empower 
knowledge workers," he explained.

"These copilots have proven to be valuable across various creative tasks, 
whether it's writing code, drafting documents, or beyond. The beauty lies in 
their collaborative nature – they offer suggestions and support, but the final 
decision and responsibility for the output rests firmly with the human user. 
This is a stark contrast to, say, a chatbot interacting with customers, where 
the AI might have a greater degree of autonomy."

And Redman said it's also not surprising that AI governance has been a common 
key initiative.

"In the face of powerful generative AI capabilities, organizations are 
naturally prioritizing risk management," he said.

"Understanding and mitigating the risks associated with each AI application is 
paramount, especially considering the growing regulatory landscape, like the EU 
AI Act. These regulations emphasize transparency and user control, ensuring 
individuals understand how their data is used and have choices in their 
interactions with AI systems." ®

--
_______________________________________________
Link mailing list
[email protected]
https://mailman.anu.edu.au/mailman/listinfo/link

Reply via email to