this post was submitted on 07 Oct 2024
113 points (100.0% liked)

Technology

37724 readers
479 users here now

A nice place to discuss rumors, happenings, innovations, and challenges in the technology sphere. We also welcome discussions on the intersections of technology and society. If it’s technological news or discussion of technology, it probably belongs here.

Remember the overriding ethos on Beehaw: Be(e) Nice. Each user you encounter here is a person, and should be treated with kindness (even if they’re wrong, or use a Linux distro you don’t like). Personal attacks will not be tolerated.

Subcommunities on Beehaw:


This community's icon was made by Aaron Schneider, under the CC-BY-NC-SA 4.0 license.

founded 2 years ago
MODERATORS
 

cross-posted from: https://lemmy.ml/post/20858435

Will AI soon surpass the human brain? If you ask employees at OpenAI, Google DeepMind and other large tech companies, it is inevitable. However, researchers at Radboud University and other institutes show new proof that those claims are overblown and unlikely to ever come to fruition. Their findings are published in Computational Brain & Behavior today.

you are viewing a single comment's thread
view the rest of the comments
[–] TranquilTurbulence@lemmy.zip 24 points 1 month ago* (last edited 1 month ago) (4 children)

Sounds really counterintuitive to say that it’s impossible.

The article says that we would run out of computing power, and that’s definitely true for current hardware and software. It’s just that they are being developed all the time, so I think we need to leave that door open. Who knows how efficient things can get within the next decade or century. The article didn’t even mention any fundamental obstacle that would make AGI completely impossible. It’s not like AGI would be violating the laws of physics.

[–] JackOverlord@beehaw.org 16 points 1 month ago (4 children)

Whenever I hear someone say that something is impossible with current technology, I think about my grandma. When she was a kid, only some important people had telephones. Doctors, police, etc.

In her lifetime we went from that to today, and, since she's still alive, even further into the future.

Whenever someone calls something impossible, I think about how far technology will progress in my own lifetime and I know that they've got no idea what they're talking about. (Unless, like you said, it's against the laws of physics. But sometimes even then I'm not so sure, cause it's not like we understand those entirely. )

[–] DdCno1@beehaw.org 13 points 1 month ago (3 children)

The thing is, we have no idea where technological progress is taking us. So far, most predictions have been wrong. 50 to 60 years ago, people thought we would already be colonizing other planets by now. Barely anyone was able to predict the Internet, smartphones, social media, etc. - the kind of technology that is actually shaping our civilization's future right now.

Another aspect that I feel is often neglected is the assumption that technological progress will continue forever or at least continue at this current rapid pace. This wasn't true in the past and we might simply be experiencing a historical anomaly right now, one that could correct itself very soon in the future, either towards stagnation or even regression.

[–] d0ntpan1c@lemmy.blahaj.zone 5 points 1 month ago (1 children)

The space example is extremely apt. Its possible we could have had tons of space stations, a moon colony, maybe even some other stuff going on around the solar system, asteroid mining, etc. But thay would have at least required the space race to continue longer and for spending to grow to create a big enoigh industry to ensure thay outcome, assuming no capacity or time issue. Alas, we took another path.

Something that seems important to us might not matter in even 10 years, or at least, not have a monetary and/or societal incentive to keep advancing.

[–] DdCno1@beehaw.org 5 points 1 month ago* (last edited 1 month ago)

I was also based on the assumption that the rapid progress of aerospace technology that happened in the 1920s to 1960s would continue onward at the same pace, whereas what actually happened was that barriers emerged that nobody was able to circumvent, like for example engineering things to withstand incredibly abrasive Moon dust (or really do anything productive on that lifeless rock), how to deal with the endless pitfalls of a long Mars journey, how to bring down the cost of launch vehicles so that grand projects like giant space stations would even be remotely possible (von Braun was already thinking about huge space stations all the way back in 1945). Many of these issues couldn't simply be solved by throwing more money at them, which is important. Deciders, both in Washington and Moscow, were smart enough to realize this in the 1970s, for the most part at least (the Space Shuttle and its Soviet clone, each a gigantic waste of money, are major counter example from this era).

The point I'm making here is that everyone assumed linear progress in this area, just like there are people currently making many billion dollar bets on linear progress in regards to computer technology in general and AI in particular, but at least, with the benefit of hindsight given past examples, there's a reasonable amount of doubt this time around.

[–] sydneybrokeit@beehaw.org 4 points 1 month ago

This wasn’t true in the past and we might simply be experiencing a historical anomaly right now

While our exact pacing might be slightly different from the pure extrapolation, human history has been a long, steady increase in the rate of invention. Access to education has meant that more people are making things, and then the next generations build on top of their work to make even bigger things.

[–] TranquilTurbulence@lemmy.zip 3 points 1 month ago

In addition, technological development can take unexpected twists and turns. For a while, it looked like analogue technology involving gears was going to solve every problem… until transistors were developed and mechanical calculators were soon forgotten. Also, the development of fertilizers revolutionized farming and and food production, which changed the world more than anyone even realized.

[–] Saganaki@lemmy.one 13 points 1 month ago (1 children)

That’s not an apt comparison.

More like “we’ll have flying cars 50 years from now.”

[–] megopie@beehaw.org 9 points 1 month ago

I love the flying car example because it reveals a huge issue with the whole “tech will get better” idea. People are still trying to make flying cars happen but it’s running in to the same fundamental issues; large things that are mechanically complex, energy intensive, and moving at high speeds in a crowded urban environments are just too expensive and dangerous.

There is no way around the physical realities, no clever trick or efficiency that will push it over some threshold of practicality.

[–] massive_bereavement@fedia.io 11 points 1 month ago (1 children)

Let's put it this way: If in our lifetime we can simulate the intelligence of a vinegar fly as general intelligence, that would be a monumental landmark in AGI. And we're far, far, far away from it.

As far as the iron age was from the metal alloys used in the Space Shuttle.

Talking about AGI simulating higher intelligence at the level of a dog or a cat, dear I say a pigeon or a crow is as far fetched as expecting ancient Egyptians to harness the power of the atom.

[–] JackOverlord@beehaw.org 6 points 1 month ago* (last edited 1 month ago) (1 children)

Let's put it this way: If in our lifetime we can simulate the intelligence of a vinegar fly as general intelligence, that would be a monumental landmark in AGI. And we're far, far, far away from it.

I get what you mean here and I agree with it, if we're talking about current "AI", which isn't anywhere close. I know, because I've programmed some simple "AIs" (Mainly ML models) myself.

But your comparison to ancient egypt is somewhat lacking, considering we had the aptly named dark ages between then and now.

Lot's of knowledge got lost all the time during humanity's history, but ever since the printing press, and more recently the internet, came into existence, this problem has all but disappeared. As long as humanity doesn't nuke itself back to said dark ages, I recon we aren't that far away from AGI, or at least something close to it. Maybe not in my lifetime, but another ~2000 years seems a little extreme.

[–] GiveMemes@jlai.lu 5 points 1 month ago
[–] TranquilTurbulence@lemmy.zip 3 points 1 month ago* (last edited 1 month ago)

Could take a while, but how long? Progress tends to be non-linear, so things can slow down and speed up suddenly. I’m pretty sure we’ll get there sooner or later unless we nuke ourselves to oblivion before that.

If AI development isn’t prioritized, it could take centuries. Maybe we’re still missing some crucial corner stores we haven’t even thought of yet. Just imagine what it was like to build an airplane in an age when the internal combustion engine hadn’t been invented yet. Maybe we’re still missing something that big. On the other hand, it could also be just around the corner, but I find it unlikely.

[–] princessnorah@lemmy.blahaj.zone 6 points 1 month ago

Actually, we do already know that we're close to a theoretical limit of increasing computing power as we currently know it. The transistor can't really get that much smaller, before it stops working.

Also, if you're talking about the article as linked, that is a mere introduction to a much longer paper.

[–] orcrist@lemm.ee 4 points 1 month ago (1 children)

The article did mention a fundamental obstacle. It said quite clearly that we would run out of resources before we had enough computing power. I suppose you could counter that by arguing that we could discover magic, or magical technology, or a lot of new resources through space exploration.

Of course things get more efficient. But in the past few decades they've gotten efficient in predictable, and mostly predicted, ways. It's certainly possible that totally unexpected things can happen. I could win the lottery next week. Is that the standard? Are you pushing the stance that says AGI is somewhat less likely than winning the lottery or getting struck by lightning, but by golly it's more than zero, how dare you suggest that it's anywhere close to zero?

[–] TranquilTurbulence@lemmy.zip 2 points 1 month ago

It really depends on your assumptions. If you assume that software and hardware will stay at the current level, then the article does present a valid point. I would argue that those assumptions are only reasonable in the short term. AGI development does depend on some big technological changes we haven’t seen yet, so it could take decades or even a century, but I wouldn’t call it impossible.

If you assumed that 1950s style vacuum tube computers were the best thing ever, you could safely say that playing a game like fortnite with your buddies living in different countries is completely impossible. Modern semiconductors and integrated circuits would have seemed pretty magical in that context.

If we assume that we’re going to be stuck with silicon, you can safely say that AGI just isn’t going to happen with these tools and methods. Since quantum computers aren’t quite useful just yet and optical computers aren’t even in the news in any meaningful way, it seems that we will be stuck with silicon for quite some time. However, in the long term, you can’t really say that for sure. Technological developments have taken sudden and unpredictable jumps from time to time.

[–] ContrarianTrail@lemm.ee 4 points 1 month ago (1 children)

The fact that human brain is capable of general intelligence tells us everything we need to know about the processing power needed to run one.

[–] emr@lemmy.sdf.org 7 points 1 month ago

Well it sets an upper bound on compute requirements at 'simulate 10^27 atoms for thirty years' remains to be seen if what we can optimize away ever converges with what's feasible to build.