this post was submitted on 20 Aug 2024
1188 points (97.8% liked)

Technology

59972 readers
2440 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 2 years ago
MODERATORS
 
  • YouTube is intensifying efforts to combat adblockers, including blocking video playback and warning users of potential account suspension.
  • Increased ads on YouTube have driven many users to adblockers, hurting both YouTube’s ad revenue and content creators reliant on ad-based income.
  • Despite these measures, many users are leaving YouTube or finding workarounds, leading creators to seek alternative revenue streams off-platform.
you are viewing a single comment's thread
view the rest of the comments
[–] UnderpantsWeevil@lemmy.world 24 points 4 months ago (3 children)

I have an LLM transcribe and summarize for me now.

Only a matter of time before LLMs start injecting their own ads into these responses.

[–] pennomi@lemmy.world 5 points 4 months ago (1 children)

Nah, local LLMs are easily in the range of transcribe/summarize. I bet you could do that nicely with llama 8B without even needing a gpu.

[–] cheese_greater@lemmy.world 2 points 4 months ago (1 children)
[–] Gigasser@lemmy.world 5 points 3 months ago

You already can I think? Ollama is something you can install, and then you can set up a webui like sillytavern for roleplays, or some other more fitting ui for whatever you want. Also, Linux is great for projects like these, on windows it's fucking a pain to set up, Linux it's easy.

[–] seaQueue@lemmy.world 3 points 4 months ago* (last edited 4 months ago)

By that point I'm pretty sure we'll have an effective compact model that can run locally and transcribe downloaded videos on reasonable hardware. Or you can just sic a paid model like chatgpt on the task. The corporate Internet is entirely focused on subscription service models now, unless you run the model yourself on local hardware you're going to end up paying someone somewhere a service fee.

Edit: y'all need to learn about minified models designed to run on edge hardware, they're a thing and often work shockingly well.

[–] Wildly_Utilize@infosec.pub 3 points 4 months ago

Local and open source