this post was submitted on 04 Sep 2024
595 points (94.5% liked)

Technology

60132 readers
2781 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 2 years ago
MODERATORS
 

Artificial intelligence is worse than humans in every way at summarising documents and might actually create additional work for people, a government trial of the technology has found.

Amazon conducted the test earlier this year for Australia’s corporate regulator the Securities and Investments Commission (ASIC) using submissions made to an inquiry. The outcome of the trial was revealed in an answer to a questions on notice at the Senate select committee on adopting artificial intelligence.

The test involved testing generative AI models before selecting one to ingest five submissions from a parliamentary inquiry into audit and consultancy firms. The most promising model, Meta’s open source model Llama2-70B, was prompted to summarise the submissions with a focus on ASIC mentions, recommendations, references to more regulation, and to include the page references and context.

Ten ASIC staff, of varying levels of seniority, were also given the same task with similar prompts. Then, a group of reviewers blindly assessed the summaries produced by both humans and AI for coherency, length, ASIC references, regulation references and for identifying recommendations. They were unaware that this exercise involved AI at all.

These reviewers overwhelmingly found that the human summaries beat out their AI competitors on every criteria and on every submission, scoring an 81% on an internal rubric compared with the machine’s 47%.

you are viewing a single comment's thread
view the rest of the comments
[–] ContrarianTrail@lemm.ee 1 points 3 months ago (1 children)

It's possible that the way of generative AI and LLMs is a dead end but that wouldn't be a stop, only a speed bump. It would only mean it takes longer for us to get there, not that we wouldn't get there.

[–] ArbitraryValue@sh.itjust.works 4 points 3 months ago (2 children)

I don't disagree, but before the recent breakthroughs I would have said that AI is like fusion power in the sense that it has been 50 years away for 50 years. If the current approach doesn't get us there, who knows how long it will take to discover one that does?

[–] kautau@lemmy.world 8 points 3 months ago (2 children)

Right and all the dogs in the race are now focused on neural networks and llms, which means for now, all the effort could be focused on a dead end. Because of the way capitalism is driving AI research, other avenues of AI research have almost effectively halted, so it will take the current AI bubble to pop before alternative research ramps up again

[–] Jesus_666@lemmy.world 4 points 3 months ago (1 children)

Like every time there's an AI bubble. And like every time changes are that in a few years public interest will wane and current generative AI will fade into the background as a technology that everyone uses but nobody cares about, just like machine translation, speech recognition, fuzzy logic, expert systems...

Even when these technologies get better with time (and machine translation certainly got a lot better since the sixties) they fail to recapture their previous levels of excitement and funding.

We currently overcome what popped the last AI bubbles by throwing an absurd amount of resources at the problem. But at some point we'll have to admit that doubling the USA's energy consumption for a year to train the next generation of LLMs in hopes of actually turning a profit this time isn't sustainable.

[–] ArbitraryValue@sh.itjust.works 2 points 3 months ago

The issue I have with referring to the current situation as a bubble is that this isn't just hype. The technology really is amazing, and far better than what people had been expecting. I do think that most current attempts to commercialize it are premature, but there's such a big first-mover advantage that it makes sense to keep losing money on attempts that are too early in order to succeed as soon as it is possible to do so.

[–] rottingleaf@lemmy.world 3 points 3 months ago* (last edited 3 months ago)

I think that's intentional. Nation states and other powers that be have working propaganda mechanisms.

A real AGI is a change most important in the sense of power, not in the sense of economy (because we know how to make new humans and educate them, it wouldn't be a qualitative change there).

All this AI gaslighting is intended to stall real advancements there.

The Web in some sense was produced in the context of AI research. In general semantic and hypertext systems were. And look what it has done to the world. They may just not want another such cataclysm.

EDIT: Also notice the shift from the hypertext paradigm to the application platform paradigm in the Web.

[–] ContrarianTrail@lemm.ee 2 points 3 months ago

The timeline doesn't really matter to me personally. As long as we accept the fact that we'll get there sooner or later it should motivate us to start thinking about the implications that comes with. Otherwise it's like knowing there's an asteroid hurling towards the earth but we'll just dismiss it by saying: "Eh, it's still 100 years away, there's no rush here"