this post was submitted on 22 Dec 2024
506 points (96.2% liked)

Technology

60073 readers
3025 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 2 years ago
MODERATORS
(page 2) 50 comments
sorted by: hot top controversial new old
[–] Zementid@feddit.nl 5 points 1 day ago

Bye bye XBox ._.

[–] ssfckdt@lemmy.blahaj.zone 29 points 1 day ago (5 children)

So you're saying we wont have any crowdsourced blockchain Web 2.0 AIs?

load more comments (5 replies)
[–] nroth@lemmy.world 56 points 1 day ago (6 children)

"Built to do my art and writing so I can do my laundry and dishes" -- Embodied agents is where the real value is. The chatbots are just fancy tech demos that folks started selling because people were buying.

[–] bradd@lemmy.world 17 points 1 day ago (5 children)

Eh, my best coworker is an LLM. Full of shit, like the rest of them, but always available and willing to help out.

load more comments (5 replies)
load more comments (5 replies)
[–] 2pt_perversion@lemmy.world 74 points 1 day ago* (last edited 1 day ago) (87 children)

There is this seeming need to discredit AI from some people that goes overboard. Some friends and family who have never really used LLMs outside of Google search feel compelled to tell me how bad it is.

But generative AIs are really good at tasks I wouldn't have imagined a computer doing just a few year ago. Even if they plateaued in place where they are right now it would lead to major shakeups in humanity's current workflow. It's not just hype.

The part that is over hyped is companies trying to jump the gun and wholesale replace workers with unproven AI substitutes. And of course the companies who try to shove AI where it doesn't really fit, like AI enabled fridges and toasters.

[–] andallthat@lemmy.world 10 points 1 day ago* (last edited 1 day ago)

Goldman Sachs, quote from the article:

“AI technology is exceptionally expensive, and to justify those costs, the technology must be able to solve complex problems, which it isn’t designed to do.”

Generative AI can indeed do impressive things from a technical standpoint, but not enough revenue has been generated so far to offset the enormous costs. Like for other technologies, It might just take time (remember how many billions Amazon burned before turning into a cash-generating machine? And Uber has also just started turning some profit) + a great deal of enshittification once more people and companies are dependent. Or it might just be a bubble.

As humans we're not great at predicting these things including of course me. My personal prediction? A few companies will make money, especially the ones that start selling AI as a service at increasingly high costs, many others will fail and both AI enthusiasts and detractors will claim they were right all along.

[–] buddascrayon@lemmy.world 70 points 1 day ago (3 children)

The part that is over hyped is companies trying to jump the gun and wholesale replace workers with unproven AI substitutes. And of course the companies who try to shove AI where it doesn't really fit, like AI enabled fridges and toasters.

This is literally the hype. This is the hype that is dying and needs to die. Because generative AI is a tool with fairly specific uses. But it is being marketed by literally everyone who has it as General AI that can "DO ALL THE THINGS!" which it's not and never will be.

load more comments (3 replies)
[–] ssfckdt@lemmy.blahaj.zone 17 points 1 day ago

This is easy to say about the output of AIs.... if you don't check their work.

Alas, checking for accuracy these days seems to be considered old fogey stuff.

load more comments (84 replies)
[–] LenielJerron@lemmy.world 134 points 2 days ago* (last edited 2 days ago) (28 children)

A big issue that a lot of these tech companies seem to have is that they don't understand what people want; they come up with an idea and then shove it into everything. There are services that I have actively stopped using because they started cramming AI into things; for example I stopped dual-booting with Windows and became Linux-only.

AI is legitimately interesting technology which definitely has specialized use-cases, e.g. sorting large amounts of data, or optimizing strategies within highly restrained circumstances (like chess or go). However, 99% of what people are pushing with AI these days as a member of the general public just seems like garbage; bad art and bad translations and incorrect answers to questions.

I do not understand all the hype around AI. I can understand the danger; people who don't see that it's bad are using it in place of people who know how to do things. But in my teaching for example I've never had any issues with students cheating using ChatGPT; I semi-regularly run the problems I assign through ChatGPT and it gets enough of them wrong that I can't imagine any student would be inclined to use ChatGPT to cheat multiple times after their grade the first time comes in. (In this sense, it's actually impressive technology - we've had computers that can do advanced math highly accurately for a while, but we've finally developed one that's worse at math than the average undergrad in a gen-ed class!)

[–] PlexSheep@infosec.pub 2 points 1 day ago

I understand some of the hype. LLMs are pretty amazing nowadays (though closedai is unethical af so don't use them).

I need to program complex cryptography code for university. Claude sonnet 3.5 solves some of the challenges instantly.

And it's not trivial stuff, but things like "how do I divide polynomials, where each coefficient of that polynomial is an element of GF(2^128)." Given the context (my source code), it adds it seamlessly, writes unit tests, and it just works. (That is important for AES-GCM, the thing TLS relies on most of the time .)

Besides that, LLMs are good at what I call moving words around. Writing cute little short stories in fictional worlds given some info material, or checking for spelling, or re-formulating a message into a very diplomatic nice message, so on.

On the other side, it's often complete BS shoehorning LLMs into things, because "AI cool word line go up".

load more comments (27 replies)
load more comments
view more: ‹ prev next ›