this post was submitted on 27 Jan 2025
883 points (98.1% liked)

Technology

61206 readers
4373 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS
 

cross-posted from: https://lemm.ee/post/53805638

you are viewing a single comment's thread
view the rest of the comments
[–] ChiefGyk3D@infosec.pub 57 points 3 days ago (4 children)

My understanding is that DeepSeek still used Nvidia just older models and way more efficiently, which was remarkable. I hope to tinker with the opensource stuff at least with a little Twitch chat bot for my streams I was already planning to do with OpenAI. Will be even more remarkable if I can run this locally.

However this is embarassing to the western companies working on AI and especially with the $500B announcement of Stargate as it proves we don't need as high end of an infrastructure to achieve the same results.

[–] sunzu2@thebrainbin.org 33 points 3 days ago

500b of trust me Bros... To shake down US taxpayer for subsidies

Read between the lines folks

[–] Dkarma@lemmy.world 5 points 3 days ago (1 children)

It's really not. This is the ai equivalent of the vc repurposing usa bombs that didn't explode when dropped.

Their model is the differentiator here but they had to figure out something more efficient in order to overcome the hardware shortcomings.

The us companies will soon outpace this by duping the model and running it on faster hw

[–] Auli@lemmy.ca 0 points 2 days ago

Throw more hardware and power at it. Build more power plants so we can use AI.

[–] Cocodapuf@lemmy.world 2 points 3 days ago (2 children)

My understanding is that DeepSeek still used Nvidia just older models

That's the funniest part here, the sell off makes no sense. So what if some companies are better at utilizing AI than others, it all runs in the same hardware. Why sell stock in the hardware company? (Besides the separate issue of it being totally overvalued at the moment)

This would be kind of like if a study showed that American pilots were more skilled than European pilots, so investors sold stock in airbus... Either way, the pilots still need planes to fly...

[–] sirboozebum@lemmy.world 15 points 3 days ago* (last edited 3 days ago) (1 children)

Perhaps the stocks were massively overvalued and any negative news was going to start this sell off regardless of its actual impact?

That is my theory anyway.

[–] Cocodapuf@lemmy.world 2 points 3 days ago

Yeah, I think that's a pretty solid theory. Makes more sense when looked at that way.

[–] hitmyspot@aussie.zone 4 points 2 days ago (1 children)

Yes, but if they already have lots of planes, they don't need to keep buying more planes. Especially if their current planes can now run for longer.

AI is not going away but it will require less computing power and less capital investment. Not entirely unexpected as a trend, but this was a rapid jump that will catch some off guard. So capital will be reallocated.

[–] Cocodapuf@lemmy.world 1 points 2 days ago (1 children)

Right, but in that metaphor, the study changes nothing, that was my point.

[–] hitmyspot@aussie.zone 1 points 1 day ago

Ok, we’ll let’s keep the plane analogy. If they could run on 50% less fuel, would you invest in airline fuel companies, thinking they will be having bumper sales figures?

[–] Corkyskog@sh.itjust.works 2 points 3 days ago (1 children)

Are there any guides to running it locally?

[–] ChiefGyk3D@infosec.pub 2 points 2 days ago

I’m using Ollama to run my LLM’s. Going to see about using it for my twitch chat bot too

https://github.com/ollama/ollama