this post was submitted on 31 May 2025
201 points (91.0% liked)

Technology

70493 readers
3312 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related news or articles.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS
all 40 comments
sorted by: hot top controversial new old
[–] KeenFlame@feddit.nu 1 points 37 minutes ago

Wonder what this has over its competitors, I hesitate to think they released this for fun though

[–] Allero@lemmy.today 1 points 1 hour ago* (last edited 1 hour ago)

There is already GPT4All.

Convenient graphical interface, any model you like (for Llama fans - of course it's there), fully local, easy to opt in or out of data collection, and no fuss to install - it's just a Linux/Windows/MacOS app.

For Linux folks, it is also available as flatpak for your convenience.

[–] cupcakezealot@lemmy.blahaj.zone 0 points 59 minutes ago

god i can't wait for the ai bubble to pop

[–] rickyrigatoni@lemm.ee 7 points 12 hours ago (1 children)

All the time I spent trying to get rid of gemini just to now download this. Am I stupid?

I wouldn't think so - it depends on your priorities.

The open source and offline nature of this without the pretenses of "Hey, we're gonna use every query you give as a data point to shove more products down your face" seems very appealing over Gemini. There's also that Gemini is constantly being shoved in our faces and preinstalled, whereas this is a completely optional download.

[–] Wazowski@lemmy.world 12 points 15 hours ago (2 children)

Excellent, I will be sure not to use this, like all Google shit.

[–] bizzle@lemmy.world 3 points 12 hours ago (1 children)

In a few years you won't be able to anyway

[–] Kolanaki@pawb.social 1 points 10 hours ago

I'm just reaching the end game faster then.

[–] RizzoTheSmall@lemm.ee 7 points 18 hours ago (1 children)

You never heard of ollama or docker model runner?

[–] fmstrat@lemmy.nowsci.com 2 points 11 hours ago

Android and iOS.

[–] toastmeister@lemmy.ca 29 points 1 day ago (3 children)

Duck.ai doesn't data mine, and has o3 mini which I have found to be very good. Its got some extra functionality like lines to break up text.

[–] Libra@lemmy.ml 3 points 17 hours ago

I've been using duck.ai recently myself and quite like it. My only complaint with it is that the chats have a length limit, so if you're working on complex projects you can run into those limits pretty quick. I use it for worldbuilding for a novel I'm working on and I have to use chatgpt for thematic stuff because it has a better memory, but otherwise it's great for quick/small things.

[–] Kuma@lemmy.world 4 points 1 day ago

Nice! I saw Mozilla also added an ai chat in the browser recently (not in the phone version that I have seen tho).

It is too bad duck.ai only runs the small models. Gpt4o-mini is not very good, it can be very inaccurate and very inconsistent :( I would like to see the 4.1-mini instead, faster and better and got function calling, so it can do web searches for example. O3 can't so it can only know what it knows until 2023.

But thanks for the information I will be looking out for when 4.1 is added!

[–] stardust@lemmy.ca 6 points 1 day ago

Yeah duck is all over bothered with since it came out since you don't even need to login to use it.

[–] AmbiguousProps@lemmy.today 36 points 1 day ago (1 children)

Why would I use this over Ollama?

[–] Greg@lemmy.ca 29 points 1 day ago (5 children)

Ollama can’t run on Android

[–] gens@programming.dev 2 points 21 hours ago

Llama.cpp (on which ollama runs on) can. And many chat programs for phones can use it.

[–] pirat@lemmy.world 2 points 21 hours ago
[–] Diplomjodler3@lemmy.world 3 points 1 day ago

Is there any useful model you can run on a phone?

[–] AmbiguousProps@lemmy.today 22 points 1 day ago (2 children)

That's fair, but I think I'd rather self host an Ollama server and connect to it with an Android client in that case. Much better performance.

[–] OhVenus_Baby@lemmy.ml 2 points 21 hours ago (1 children)

How is Ollama compared to GPT models? I used the paid tier for work and I'm curious how this stacks up.

[–] AmbiguousProps@lemmy.today 1 points 19 hours ago

It's decent, with the deepseek model anyway. It's not as fast and has a lower parameter count though. You might just need to try it and see if it fits your needs or not.

[–] Greg@lemmy.ca 4 points 1 day ago

Yes, that's my setup. But this will be useful for cases where internet connection is not reliable

[–] Euphoma@lemmy.ml 6 points 1 day ago (1 children)
[–] Greg@lemmy.ca 3 points 1 day ago (1 children)

Has this actually been done? If so, I assume it would only be able to use the CPU

[–] Euphoma@lemmy.ml 6 points 1 day ago

Yeah I have it in termux. Ollama is in the package repos for termux. The speed it generates does feel like cpu speed but idk

[–] rirus@feddit.org 23 points 1 day ago

Alibaba also provides an OpenSource App, it even has support for their multimodal voice chat Model qwen2.5 omni: https://github.com/alibaba/MNN

[–] NGC2346@sh.itjust.works 7 points 1 day ago (1 children)

Enclave on iOS does the trick for the rare times i need a local LLM

[–] th3dogcow@lemmy.world 4 points 1 day ago

Didn't know about this. Checking it out now, thanks!

[–] moonlight6205@lemm.ee 9 points 1 day ago (2 children)
[–] Rhaedas@fedia.io 44 points 1 day ago (2 children)

And unmonitored? Don't trust anything from Google anymore.

What makes this better than Ollama?

[–] Angelusz@lemmy.world 23 points 1 day ago (1 children)

Quote: "all running locally, without needing an internet connection once the model is loaded. Experiment with different models, chat, ask questions with images, explore prompts, and more!"

So you can download it and set the device to airplane mode, never go online again - they won't be able to monitor anything, even if there's code for that included.

[–] taladar@sh.itjust.works 11 points 1 day ago

That is exactly what Ollama does too.

[–] GreenKnight23@lemmy.world 6 points 1 day ago (1 children)

everything is unmonitored if you don't connect to the network.

[–] flightyhobler@lemmy.world 5 points 1 day ago (1 children)

But not everything works in those conditions.

[–] GreenKnight23@lemmy.world 1 points 1 day ago (1 children)

it does if you make it work in those conditions.

software that "phones home" is easy to fool.

[–] Deckname@discuss.tchncs.de 2 points 1 day ago (1 children)

Just firewall the software or is there anything more fancy i would need to do?

[–] GreenKnight23@lemmy.world 1 points 20 hours ago

typically the phone home is looking for a response to unlock.

use a packet sniffer to see what the request/response is and replicate it with a proxy or response server.

this is also know as a man-in-the-middle (mitm).

takes skill and knowledge to do, but once you do a few dozen it's pretty easy since most software "phone homes" are looking for static non-encrypted responses.

[–] ofcourse@lemmy.ml 5 points 1 day ago

Censoring is model dependent so you can select one of the models without the guardrails.