this post was submitted on 10 Mar 2024
42 points (100.0% liked)
Comradeship // Freechat
2168 readers
127 users here now
Talk about whatever, respecting the rules established by Lemmygrad. Failing to comply with the rules will grant you a few warnings, insisting on breaking them will grant you a beautiful shiny banwall.
A community for comrades to chat and talk about whatever doesn't fit other communities
founded 3 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
I read about this on the cursed orange site. Some guy talked about going on a walk with his wireless warplugs on, talking to ChatGPT's audio interface discussing some world building he was doing.
Are there any LLM services that can be reasonably used without paying? I tried some llamafiles but seems like my laptop cannot handle them well.
As long as you don't care about your inputs being harvested, gemini is free currently. I've been using GPT4All to run stuff locally, but if your laptop is having trouble with llamafiles, then it's probably gonna have trouble with that too.
What model do you run?
I find I like Wizard 1.2 and Hermes the best
On the topic of GPT4ALL, I'm curious is there an equivalent of that that but for txt2img/img2img models? All the FOSS txt2img stuff I've tried so far is either buggy (some of the projects I tried don't even compile), require a stupid amount of third party dependencies, are made with NVidia hardware in mind while everyone else is second class or require unspeakable amounts of VRAM.
automatic1111 webui launcher, its stable diffusion. fun fact its icon is a pic of ho chi minh
if you wait, stable diffusion 3 is coming out soon. nvidia will run faster because its tensors are better unfortunately. SD is more ethical than others, you can load up models that are trained only on public art and pics
I'm pretty sure I tried that one but it kept running out of VRAM. Also it utilizes proprietary AMD/NVidia software stacks which are a pain to set up. GPT4ALL is a lot better in that regard, they just use Vulkan compute shaders to run the models.
could try out the turbo models, might help
There's also ComfyUI, but the learning curve is a bit steeper https://github.com/comfyanonymous/ComfyUI
although there's CushyStudio frontend for it that's more user friendly https://github.com/rvion/CushyStudio
ComfyUI seems like the most promising but it also uses ROCm/CUDA which don't officially support any of my current GPUs (models load successfully but midway through computing it fails). Why can't everyone just use compute shaders lol.
Oh yeah that whole thing is just such a mess, another L for proprietary tech.
seconding gpt4all, makes it quick and easy to run and if youre fancy you can stream the output from your computer to your phone. i run a capybara-hermes-mistral mix but i would suggest starting with mistral instruct until claude3 comes out