schizo

joined 5 months ago
[–] schizo@forum.uncomfortable.business 4 points 1 hour ago (1 children)

Privacy regulations are all fine and dandy, but even with the strictest ones in place,

They're also subject to interpretation, regulatory capture, as well as just plain being ignored when it's sufficiently convenient for the regulators to do so.

"There ought to be a law!" is nice, but it's not a solution when there's a good couple of centuries of modern regulatory frameworks having had existed, and a couple centuries of endless examples of where absolutely none of it matters when sufficient money and power is in play.

Like, for example, the GDPR: it made a lot of shit illegal under penalty of company-breaking penalties.

So uh, nobody in the EU has had their personal data misused since it was passed? And all the big data brokers that are violating it have been fined out of business?

And this is, of course, ignoring the itty bitty little fact that you have to be aware of the misuse of the data: if some dude does some shady shit quietly, then well, nobody knows it happened to even bring action?

How exactly are “communities offering services” a different thing than “hosted software”?

I think what they're saying is that the ideal wouldn't be to force everyone to host their own, but rather for the people who want to run stuff to offer them to their friends and family.

Kinda like how your mechanic neighbor sometimes helps you do shit on your car: one person shares a skill they have, and the other person also benefits. And then later your neighbor will ask you to babysit their kids, and shit.

Basically: a very very goofy way of saying "Hey! Do nice things for your friends and family, because that's kinda how life used to work."

[–] schizo@forum.uncomfortable.business 2 points 17 hours ago (2 children)

AI model of that type is safe to deploy anywhere

Yeah, I think you've made a mistake in thinking that this is going to be usable as generative AI.

I'd bet $5 this is just a fancy machine learning algorithm that takes a submitted image, does machine learning nonsense with it, and returns a 'there is a high probability this is an illicit image of a child', and not something you could use to actually generate CSAM with.

You want something that's capable of assessing the similarities between a submitted image and a group of known bad images, but that doesn't mean the dataset is in any way usable for anything other than that one specific task - AI/ML in use cases like this is super broad and has been a thing for decades before the whole 'AI == generative AI' thing became what everyone is thinking.

But, in any case: the PhotoDNA database is in one place and access to it is scaled by the merit of uh, lots of money?

And of course, any 'unscrupulous engineer' that may have any plans for doing anything with this is probably not a complete idiot, even if a pedo: they're going to have shockingly good access controls and logging and well, if you're in the US, if the dude takes this database and generates a couple of CSAM images using it, the penalty is, for most people, spending the rest of their life in prison.

Feds don't fuck around with creation or distribution charges.

[–] schizo@forum.uncomfortable.business 4 points 17 hours ago (2 children)

I assume the KDE implementation resizes to default when you stop shaking it.

I could totally see someone coding a function that increases the mouse pointer by x% every y mouse shakes, and then neglecting to put in a size cap.

[–] schizo@forum.uncomfortable.business 4 points 17 hours ago (3 children)

This is kinda old information, but my understanding was that there were 3 issues with dasiy-chained UPSes.

The first is that you're potentially going to cause a ground loop, which is not healthy for the life of anything plugged into those UPSes.

The second is that there's a potential for a voltage droop going through from the first to second UPS, which means the UPSes will flap constantly and screw their batteries up, though I'd be shocked if that was necessarily still true for modern high-quality units.

And of course, the UPS itself won't be outputting a proper sinewave when it's on battery, which means your 2nd UPS in the chain will freak out (though again, maybe modern ones don't have that limitation).

[–] schizo@forum.uncomfortable.business 13 points 18 hours ago (5 children)

Yeah, and Windows and OS X both do it as well.

Though there being no upper limit to the size is amusing.

[–] schizo@forum.uncomfortable.business 2 points 18 hours ago (4 children)

comparative scale of the content involved

PhotoDNA is based on image hashes, as well as some magic that works on partial hashes: resizing the image, or changing the focus point, or fiddling with the color depth or whatever won't break a PhotoDNA identification.

But, of course, that means for PhotoDNA to be useful, the training set is literally 'every CSAM image in existance', so it's not really like you're training on a lot less data than an AI model would want or need.

The big safeguard, such as it is, is that you basically only query an API with an image and it tells you if PhotoDNA has it in the database, so there's no chance of the training data being shared.

Of course, there's also no reason you can't do that with an AI model, either, and I'd be shocked if that's not exactly how they've configured it.

[–] schizo@forum.uncomfortable.business 4 points 20 hours ago (1 children)

The problem I ran into is that every single platform that primarily interacted with Mastodon (The keys, etc.) had the same exact same set of problems.

While yes, my Firefish instance had search, what was it searching? Local data only, and once I figured out that Mastodon-style replies didn't federate to all of someone's followers, it became pretty clear that it was uh, not very useful.

You can search, but any given server may or may not have access to data you actually want and thus, well, you just plain cannot meaningfully search for shit unless you go to one of the mega instances, or join giant piles of relays and store gigabyte upon gigabyte upon gigabyte of garbage data you do not care about.

The whole implementation is kinda garbage for search-based discovery from it's very basic design all the way through to everyone's implementations.

[–] schizo@forum.uncomfortable.business 4 points 20 hours ago (6 children)

first time law enforcement are sharing actual csam with a technology company

It's very much not: PhotoDNA, which is/was the gold standard for content identification, is a collaboration between a whole bunch of LEOs and Microsoft. The end user is only going to get a 'yes/no idea' result on a matched hash, but that database was built on real content working with Microsoft.

Disclaimer: below is my experience dealing with this shit from ~2015-2020, so ymmv, take it with some salt, etc.

Law enforcement is also rarely the first-responder to these issues, either: in the US, at least, reports will come to the hosting/service provider first for validation and THEN to NCMEC and LEOs, if the hosting provider confirms what the content is. Even reports that are sent from NCMEC to the provider aren't being handled by law enforcement as the first step, usually.

And as for validating reports, that's done by looking at it without all the 'access controls and safeguards' you think there are, other than a very thin layer of CYA on the part of the company involved. You get a report, and once PhotoDNA says 'no fucking clue, you figure it out' (which, IME, was basically 90% of the time) a human is going to look at it and make a determination, and then file a report with NCMEC or whatever, if it turns out to be CSAM.

Frankly, after having done that for far too fucking long, if this AI tool can reduce the amount of horrible shit someone doing the reviews has to look at, I'm 100% for it.

CSAM is (grossly) a big business, and the 'new content' funnel is fucking enormous and is why an extremely delayed and reactive thing like PhotoDNA isn't all that effective is that, well, there's a fuckload of children being abused and a fuckload of abusers escaping being caught simply because there's too much shit to look at and handle effectively and thus any response to anything is super super slow.

This looks like a solution to make it so less people have to be involved in validation, and could be damn near instant in responding to suspected material that does need validation, which will do a good job of at least pushing the shit out of easy (ier?) availability and out of more public spaces, which honestly, is probably the best thing that is going to be managed unless the countries producing this shit start caring and going after the producers which I'm not holding my breath on.

[–] schizo@forum.uncomfortable.business 14 points 22 hours ago (4 children)

For me, it's full text search.

I tend to want to find an opinion on something very specific, so if I can just toss a phrase or model number or name of something into a search field and get actual non-AI, non-advertisement, non-stupid-shit results, that'd be absolutely ideal.

Like, say, how Google worked 15 years ago.

[–] schizo@forum.uncomfortable.business 7 points 22 hours ago (8 children)

First: you'd probably be shocked how many pedos have zero opsec and just post shit/upload shit in the plain.

By which I mean most of them, because those pieces of crap don't know shit about shit and don't encrypt anything and just assume crap is private.

And second, yeah, I'll catch kids generating CSAM, but it'll catch everyone else too, so that's probably a fair trade.

 

Made this mostly because I've found putting RSS feeds into Lemmy useful since my doom-scrolling has reduced to just Lemmy and figured I'm probably not the only person that'd find this useful.

It's pulling 6 RSS feeds that provide free games for Steam, Gog, Epic, and Humble.

Nothing shockingly world-changing, but hey, free games.

!freegames@forum.uncomfortable.business

 

So I'm looking for a laptop, but before you downvote and move on, I've got a twist: I'm looking for a laptop with Linux support that's going to intentionally be console-only and rely on TUIs to make a lower-distraction device.

I was looking at older Thinkpads with 4:3 screens and the good keyboard before Lenovo went all chicklet with them, but I'm kinda concluding they're both way too expensive AND way too old to be a reasonable choice at this point.

A X220 or T40-whatever would be great and be the perfect aesthetic, but they're expensive, hard to find parts for, and using enough crusty old shit that this becomes yet another delve into retro computing and not one into practical, useful computing which is the goal here.

So, anyone have any recommendations of any devices in the last decade that have a reasonable keyboard, screen, use modern enough components that you can source new drives and RAM and batteries and such, and preferably aren't coated in a coating that's going to turn to sticky goo?

Thin(ner) and light(er) would be nice, but probably not a dealbreaker if the rest of the pieces align. This will be almost entirely used at a table for writing and such.

 

Basically, the court said that algorithmically selected content doesn't qualify for Section 230 protections, which could be a massive impact to every social media platform out there that has any sort of algorithm selecting content, which, well, is all of them.

Definitely something that's going to be interesting watching play out.

 

I have a question for the hive mind: what is the point of this, exactly?

I mean, I understand the attempt to gain access, and I understand why 2fa codes can be valuable to attempt to phish but that's like, not the thing here.

They just spam dozens to hundreds of these (I'm showing over 400 in my inbox right now) but like, even if I WANTED to give these codes to the attacker, I have no damn clue who the dude in China that's doing this is.

I'm confused as to what they hope to gain by trying over and over and over every couple of hours because it feels like there's no upside to whomever is running this bot, but I probably have missed a memo on some TTP around this, heh.

 

So I've got a home server that's having issues with services flapping and I'm trying to figure out what toolchain would be actually useful for telling me why it's happening, and not just when it happened.

Using UptimeKuma, and it's happy enough to tell me that it couldn't connect or a 503 happened or whatever, but that's kinda useless because the service is essentially immediately working by the time I get the notice.

What tooling would be a little more detailed in to the why, so I can determine the fault and fix it?

I'm not sure if it's the ISP, something in my networking configuration, something on the home server, a bad cable, or whatever because I see nothing in logs related to the application or the underlying host that would indicate anything even happened.

It's also not EVERY service on the server at once, but rather just one or two while the other pile doesn't alert.

In sort: it's annoying and I'm not really making headway for something that can do a better job at root-cause-ing what's going on.

 

Just got an email thanking me for being a 5-node/free user, but Portainer isn't free and I need to stop being a cheap-ass and pay them because blah blah economic times enshittification blah blah blah.

I've moved off them a while ago, but figured I'd see if they emailed EVERYONE about this?

A good time to ditch them if you haven't, I suppose.

 

I'm wanting to add a bunch of energy monitoring stuff so I can both track costs, and maybe implement automation to turn stuff on and off based on power costs and timing.

I'm using some TPlink based plugs right now which are like, fine, but I'm wanting to add something like 6 to 10 more monitoring devices/relays.

Anyone have experience with a bunch of shelly devices and if there's any weird behavior I should be aware of?

Assume I have good enough wifi to handle adding another 10 devices to it, but beyond that any gotchas?

view more: next ›