this post was submitted on 06 Jul 2024
1216 points (99.2% liked)
Microblog Memes
5710 readers
4034 users here now
A place to share screenshots of Microblog posts, whether from Mastodon, tumblr, ~~Twitter~~ X, KBin, Threads or elsewhere.
Created as an evolution of White People Twitter and other tweet-capture subreddits.
Rules:
- Please put at least one word relevant to the post in the post title.
- Be nice.
- No advertising, brand promotion or guerilla marketing.
- Posters are encouraged to link to the toot or tweet etc in the description of posts.
Related communities:
founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
No, AGI will destroy the world because it doesn't care about our moral values (such as keeping us alive), and an instrumental goal of most goals it could be programmed with is killing us.
https://en.wikipedia.org/wiki/Instrumental_convergence
Agi won't kill us because it doesn't exist and won't emerge from what we call ai now.
We can also just unplug it if necessary.
True, but do you really think the investors* will allow the companies behind the AI or running the servers the AI are hosted on would allow that?
*parasites in human clothes
You clearly didn't read the Wikipedia article I linked; an intelligent AGI would not let you unplug it.
And regardless of whether it emerges from current AI, or is developed in a totally different way, there is no reason besides blind optimism (ie, burying your head in the sand) to feel certain it will never exist.
I am not saying it will never exist. I am saying it doesn't exist right now and doesn't look like it will for a long time. We clearly have way more pressing matters to worry about, like climate change for example.
No reason to assume it will take a very long time to happen. Its best to take it as a serious threat, unless you want extremely rapid climate change.
Also, you didn't address the fact that my comment addressed the second part of your original comment too. Do you accept the correction?
No reason to assume it will take a very long time for ragnarok to happen either. Better prepare now!
If you give the ai enough power, sure, it won't let you unplug it maybe. Still don't really see how it wants to prevent a hardware killswitch from being activated except for guarding it or disabling it somehow.
Why would you put a hardware kill switch on a military robot? Then the enemy can just switch it off while it’s killing then.
Why put a brain in a military robot? It'd just have a higher chance to fuck you over.
There is also no proof that any form of agi is on the way or even possible. Preparing for it over any of the things we do have proof of makes about as much sense as prepping for a zombie apocalypse.