this post was submitted on 06 Jul 2024
142 points (90.3% liked)

Technology

59429 readers
3279 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] Sgagvefey@lemmynsfw.com 28 points 4 months ago (26 children)

Though it sounds extreme, there are a lot of smart people in the AI community who truly believe AI could end humanity.

No. There are not.

Believing anything resembling current tools has the capacity to end humanity in incontrovertible proof that you are not smart.

[–] 5C5C5C@programming.dev 32 points 4 months ago (23 children)

AI person reporting in. Without saying whether or not I personally believe that the current tools will lead to the end of humanity, I'll point out a few possibilities that I find concerning about what's going on:

  • The hype around AI is being used to justify mass layoffs, where humans are being replaced by tools that do a questionable job and can't really understand the things those humans could understand. Whether or not the AI can do as good of a job according to some statical measurement is less relevant than the fact that a human is less likely to make an extremely grave mistake and more likely to be able to recognize when that does happen. I'm concerned this will lead to cross-industry enshitification on an unprecedented scale.

  • The foundation models consume a huge amount of energy. The more impressive you want it to be, the more energy it needs. As long as the data centers which run them are dependent on fossil fuels, they'll be pumping a huge amount of carbon in the air just to do replace jobs that we didn't need to have replaced.

  • As these tools are used more and more, they're going to end up "learning" from content created by themselves instead of something that's closer to a ground truth. It's hard to predict what kind of degradation of service will come from this, but the more we create systems that rely on these tools, the more harm it will do to us.

  • Given the cost and nature of these tools, they're likely to yield the most benefit to moneyed interests that want to automate the systems that maintain their power and wealth. E.g. generating large amounts of convincing disinformation to manipulate the public into supporting politicians or policies that benefit a small number of wealthy people in the short term while locking humanity into a path towards destruction.

And none of this accounts for possible future iterations of AI tools that may be far more capable than what exists today. That future technology will most likely be controlled by powerful people who are primarily interested in using it to bolster the systems that keep them in power, to the detriment of humanity as a whole.

Personally I'm far less concerned about a malicious AI intentionally doing harm to humanity than AI being used as a weapon by unscrupulous people.

[–] Sgagvefey@lemmynsfw.com 3 points 4 months ago (2 children)

Zero of these things are impacted by this legislation in any way.

This is exclusively the mentally unstable "killer AI" nonsense. We're not even 1% of 1% of the way to anything resembling agency.

[–] ShittyBeatlesFCPres@lemmy.world 2 points 4 months ago (1 children)

It’s good for marketing, though. “Ah, our software is so powerful, it could destroy humanity! Please pass a bill saying so while we market friendly chatbots to the public while actually making money by selling our products to despots and warmongers that might actually end humanity.”

[–] Sgagvefey@lemmynsfw.com 3 points 4 months ago

It's regulatory capture. Add deluded barriers to entry to make it difficult for competition and community projects to develop, and you have a monopoly.

[–] 5C5C5C@programming.dev 0 points 4 months ago (1 children)

Sure, but this outcome is not at all surprising. There are plenty of smart AI people that have nuanced views of what kind of threat could be posed by recklessly unleashing tools that we don't fully understand into the hands of people who are likely to do harmful things with them.

It's not surprising that those valid nuanced concerns get translated into overly simplistic misrepresentations entangled with pop sci fi panic around rogue AI as they try to move into public discourse.

[–] Sgagvefey@lemmynsfw.com 2 points 4 months ago* (last edited 4 months ago) (1 children)

We do fully understand them. Not knowing the exact reason they come to a model doesn't mean the algorithm has a shred of mystery involved. It's like saying we don't understand fluid dynamics because it's computationally heavy.

It's autocomplete with a really big training set and a really big model. It cannot possibly develop agency. It's hundreds of orders of magnitude of complexity short of a human.

[–] 5C5C5C@programming.dev 0 points 4 months ago (1 children)

That's not what an algorithms researcher means when we talk about "understanding". Obviously we know the mechanism by which it operates, it's not an unknown alien technology that dropped into our laps.

Understanding an algorithm means being able to predict the characteristics of its outputs based on the characteristics of its inputs. E.g. will it give an optimal solution to a problem that we pose? Will its response satisfy certain constraints or fall within certain bounds?

Figuring this stuff out for foundation models is an active area of research, and the absence of this predictability is an enormous safety concern for any use cases where the output can be consequential.

It cannot possibly develop agency.

I don't believe I've suggested anywhere that I think it will, but I'll play around with this concern anyway... There's a lot of discussion going on about having models feed back on themselves to learn from their own output. I don't find it all that hard to imagine that something we could reasonably consider self awareness could be formed by a very complex neural network that is able to consume and process its own outputs. And once self awareness starts to form, it's not that hard for me to imagine a sense of agency following. I have no idea what the model might use that agency for, but I don't think it's all that far fetched to consider the possibility of it happening.

[–] Sgagvefey@lemmynsfw.com 1 points 4 months ago (1 children)

There are plenty of nondeterministic algorithms. It's not a special trait. There are plenty of algorithms with actual emergent behavior, which LLMs don't have to any meaningful extent. We absolutely understand how LLMs work

The answer to both of your questions is not some unsolved mystery. It's "of course not". That's not what they do and fundamentally requires a much more complex architecture to even approach.

[–] 5C5C5C@programming.dev 0 points 4 months ago (1 children)

Non-deterministic algorithms such as Monte Carlo methods or simulated annealing can still be constrained to an acceptable state space. How to do this effectively for LLMs is a very open question, largely because the state space of the problems that they are applied to is incomprehensibly huge.

[–] Sgagvefey@lemmynsfw.com 1 points 4 months ago (1 children)

It's only an "open question" if you are somehow confused by the fact that it's a super simple algorithm that cannot ever possibly be used like that.

It may be a small part of a proper architecture for a functional solution, but there's no possibility that it will ever be doing the heavy lifting. It is what it is, and that's an obvious dead end.

[–] 5C5C5C@programming.dev 0 points 4 months ago* (last edited 4 months ago) (1 children)

Literally nothing you've said gives any indication that you actually know the current state of foundation model research. I won't claim it's my research specialty, but I work directly with people whose full time job is research and tuning on foundation models, and everything I'm saying is being relayed from conversations that I have with them.

"Cannot ever possibly be used like that".. Like what specifically? To drive a car? That's being done. To give financial advice? That's being done. To console people who are suicidal or at risk of harming themselves? That's being done. To make kill / no kill decisions in an active warzone? It's being considered (if not already being done in secret).

This technology is being used in extremely consequential positions despite having very weak guarantees around safety. This should give any reasonable person pause. I'm not taking any firm stance on whether this specific regulation is the right approach, but if you think there should be no legal accountability for the outcomes of how this technology gets used then I guess you're someone who thinks seatbelts should be optional in cars and it's okay for airplanes to fall out of the sky due to neglect.

[–] Sgagvefey@lemmynsfw.com 1 points 4 months ago (1 children)

For anything where you would ever expect a predictable, useful outcome to an arbitrary input. There is no possible path to LLMs ever doing anything close to that.

LLMs aren't driving cars. LLMs aren't doing financial modeling. Those are entirely different tools with heavily hand crafted models to specific applications.

Anyone using an LLM to provide therapy should get multiple life sentences in prison regardless of outcomes. There is no possible way to LLMs ever being actually useful for therapy. It's just a random text generator that's tuned well enough to sound good. It has no substance and the underlying tech cannot possibly develop substance.

[–] 5C5C5C@programming.dev 0 points 4 months ago* (last edited 4 months ago)

I can't tell if you're suggesting that foundation models (which is the underpinning technology of LLMs) aren't being used for the things that I said they're being used for, but I can assure you they are, either in commercial R&D or in live commercial products.

The fact that they shouldn't be used for these things is something we can certainly agree on, but the fact remains that they are.

Sources:

So this all goes back to my point that some form of accountability is needed for how these tools get used. I haven't examined the specific legislation proposal enough to give any firm opinion on it, but I think it's a good thing that the conversation is happening in a serious way.

load more comments (20 replies)
load more comments (22 replies)