this post was submitted on 27 Jan 2025
883 points (98.1% liked)

Technology

61206 readers
4373 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS
 

cross-posted from: https://lemm.ee/post/53805638

you are viewing a single comment's thread
view the rest of the comments
[–] thespcicifcocean@lemmy.world 11 points 2 days ago (2 children)

nvidia falling doesn't make much sense to me, GPUs are still needed to run the model. Unless Nvidia is involved in its own AI model or something?

[–] Jhex@lemmy.world 20 points 2 days ago (1 children)

DeepSeek proved you didn't need anywhere near as much hardware to train or run an even better AI model

Imagine what would happen to oil prices if a manufacturer comes out with a full ice car that can run 1000 miles per gallon... Instead of the standard American 3 miles per 1.5 gallons hehehe

[–] HoMaster@lemm.ee 13 points 2 days ago (1 children)

https://en.wikipedia.org/wiki/Jevons_paradox

more efficient use of oil will lead to increased demand, and will not slow the arrival or the effects of peak oil.

Energy demand is infinite and so is the demand for computing power because humans always want to do MORE.

[–] Jhex@lemmy.world 1 points 2 days ago* (last edited 2 days ago)

Yes but that's not the point... If you can buy a house for $1000 nobody would buy a similar house for $500000

Eventually the field would even out and maybe demand would surpass current levels, but for the time being, Nvidia's offer seem to be a giant surplus and speculators will speculate

[–] racemaniac@lemmy.dbzer0.com 24 points 2 days ago (1 children)

If you need far less computing power to train the models, far less gpus are needed, and that hurts nvidia

[–] thespcicifcocean@lemmy.world 6 points 2 days ago (1 children)

does it really need less power? I'm playing around with it now and I'm pretty impressed so far. it can do math, at least.

[–] racemaniac@lemmy.dbzer0.com 18 points 2 days ago (2 children)

That's the claim, it has apparently been trained using a fraction of the compute power of the GPT models and achieves similar results.

[–] thespcicifcocean@lemmy.world 7 points 2 days ago (2 children)

fascinating. my boss really bought into the tech bro bullshit, every time we get coffee as a team, he's always going on and on about how chatGPT will be the savior of humanity, increase productivity so much that we'll have a 2 day work week, blah blah blah.

I've been on his shit list lately because i had to take some medical leave and didn't deliver my project on time.

Now that this thing is open sourced, I can bring it to him, tell him it out performs even chatgpt O1 or whatever it is, and tell him that we can operate it locally. I'll be off the shit list and back into his good graces and maybe even get a raise.

[–] Zink@programming.dev 5 points 2 days ago

Your boss sounds like he buys into bullshit for a living. Maybe that’s what drew him to the job, lol.

I think believing in our corporate AI overlords is even overshadowed by believing those same corporations would pass the productivity gains on to their employees.

[–] MutilationWave@lemmy.world 1 points 2 days ago

Make sure you explain open source to them so they know it's not spyware.

[–] HereIAm@lemmy.world 4 points 2 days ago (1 children)

But I feel like that will just lead to more training with the same (or more) hardware with a more efficient model. Bitcoin mining didn't slow down only because it got harder. However I don't know enough about the training process. I assume more efficient use of the hardware would allow for larger models to be trained on the same hardware and training data?

[–] racemaniac@lemmy.dbzer0.com 3 points 2 days ago

They'll probably do that, but that's assuming we aren't past the point of diminishing returns.

The current LLM's are pretty basic in how they work, and it could be that with the current training we're near what they'll ever be capable of. They'll of course invest a billion in training a new generation, but if it's only marginally better than the current one, they won't keep investing billions into it if it doesn't really improve the results.