this post was submitted on 09 Jun 2025
814 points (92.0% liked)

Technology

71266 readers
5302 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related news or articles.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS
top 50 comments
sorted by: hot top controversial new old
[–] untakenusername@sh.itjust.works 1 points 2 hours ago

this is because an LLM is not made for playing chess

[–] stevedice@sh.itjust.works 7 points 8 hours ago

2025 Mazda MX-5 Miata 'got absolutely wrecked' by Inflatable Boat in beginner's boat racing match — Mazda's newest model bamboozled by 1930s technology.

[–] FourWaveforms@lemm.ee 5 points 10 hours ago

If you don't play chess, the Atari is probably going to beat you as well.

LLMs are only good at things to the extent that they have been well-trained in the relevant areas. Not just learning to predict text string sequences, but reinforcement learning after that, where a human or some other agent says "this answer is better than that one" enough times in enough of the right contexts. It mimics the way humans learn, which is through repeated and diverse exposure.

If they set up a system to train it against some chess program, or (much simpler) simply gave it a tool call, it would do much better. Tool calling already exists and would be by far the easiest way.

It could also be instructed to write a chess solver program and then run it, at which point it would be on par with the Atari, but it wouldn't compete well with a serious chess solver.

[–] NeilBru@lemmy.world 76 points 1 day ago* (last edited 1 day ago) (4 children)

An LLM is a poor computational/predictive paradigm for playing chess.

[–] surph_ninja@lemmy.world 30 points 1 day ago (1 children)

This just in: a hammer makes a poor screwdriver.

[–] WhyJiffie@sh.itjust.works 8 points 1 day ago

LLMs are more like a leaf blower though

[–] Takapapatapaka@lemmy.world 12 points 1 day ago (4 children)

Actually, a very specific model (chatgpt3.5-turbo-instruct) was pretty good at chess (around 1700 elo if i remember correctly).

load more comments (4 replies)
[–] Bleys@lemmy.world 5 points 1 day ago (1 children)

The underlying neural network tech is the same as what the best chess AIs (AlphaZero, Leela) use. The problem is, as you said, that ChatGPT is designed specifically as an LLM so it’s been optimized strictly to write semi-coherent text first, and then any problem solving beyond that is ancillary. Which should say a lot about how inconsistent ChatGPT is at solving problems, given that it’s not actually optimized for any specific use cases.

[–] NeilBru@lemmy.world 3 points 1 day ago* (last edited 21 hours ago)

Yes, I agree wholeheartedly with your clarification.

My career path, as I stated in a different comment in regards to neural networks, is focused on generative DNNs for CAD applications and parametric 3D modeling. Before that, I began as a researcher in cancerous tissue classification and object detection in medical diagnostic imaging.

Thus, large language models are well out of my area of expertise in terms of the architecture of their models.

However, fundamentally it boils down to the fact that the specific large language model used was designed to predict text and not necessarily solve problems/play games to "win"/"survive".

(I admit that I'm just parroting what you stated and maybe rehashing what I stated even before that, but I like repeating and refining in simple terms to practice explaining to laymen and, dare I say, clients. It helps me feel as if I don't come off too pompously when talking about this subject to others; forgive my tedium.)

load more comments (1 replies)
[–] jsomae@lemmy.ml 13 points 1 day ago (1 children)

Using an LLM as a chess engine is like using a power tool as a table leg. Pretty funny honestly, but it's obviously not going to be good at it, at least not without scaffolding.

[–] kent_eh@lemmy.ca 3 points 12 hours ago (1 children)

is like using a power tool as a table leg.

Then again, our corporate lords and masters are trying to replace all manner of skilled workers with those same LLM "AI" tools.

And clearly that will backfire on them and they'll eventually scramble to find people with the needed skills, but in the meantime tons of people will have lost their source of income.

[–] jsomae@lemmy.ml 1 points 9 hours ago* (last edited 9 hours ago)

If you believe LLMs are not good at anything then there should be relatively little to worry about in the long-term, but I am more concerned.

It's not obvious to me that it will backfire for them, because I believe LLMs are good at some things (that is, when they are used correctly, for the correct tasks). Currently they're being applied to far more use cases than they are likely to be good at -- either because they're overhyped or our corporate lords and masters are just experimenting to find out what they're good at and what not. Some of these cases will be like chess, but others will be like code*.

(* not saying LLMs are good at code in general, but for some coding applications I believe they are vastly more efficient than humans, even if a human expert can currently write higher-quality less-buggy code.)

[–] cley_faye@lemmy.world 25 points 1 day ago

Ah, you used logic. That's the issue. They don't do that.

[–] nednobbins@lemm.ee 50 points 1 day ago (2 children)

Sometimes it seems like most of these AI articles are written by AIs with bad prompts.

Human journalists would hopefully do a little research. A quick search would reveal that researches have been publishing about this for over a year so there's no need to sensationalize it. Perhaps the human journalist could have spent a little time talking about why LLMs are bad at chess and how researchers are approaching the problem.

LLMs on the other hand, are very good at producing clickbait articles with low information content.

[–] nova_ad_vitum@lemmy.ca 24 points 1 day ago (7 children)

Gotham chess has a video of making chatgpt play chess against stockfish. Spoiler: chatgpt does not do well. It plays okay for a few moves but then the moment it gets in trouble it straight up cheats. Telling it to follow the rules of chess doesn't help.

This sort of gets to the heart of LLM-based "AI". That one example to me really shows that there's no actual reasoning happening inside. It's producing answers that statistically look like answers that might be given based on that input.

For some things it even works. But calling this intelligence is dubious at best.

[–] Ultraviolet@lemmy.world 6 points 1 day ago* (last edited 1 day ago)

Because it doesn't have any understanding of the rules of chess or even an internal model of the game state, it just has the text of chess games in its training data and can reproduce the notation, but nothing to prevent it from making illegal moves, trying to move or capture pieces that don't exist, incorrectly declaring check/checkmate, or any number of nonsensical things.

load more comments (6 replies)
[–] LovableSidekick@lemmy.world 6 points 1 day ago* (last edited 1 day ago) (3 children)

In this case it's not even bad prompts, it's a problem domain ChatGPT wasn't designed to be good at. It's like saying modern medicine is clearly bullshit because a doctor loses a basketball game.

load more comments (3 replies)
[–] Halosheep@lemm.ee 43 points 1 day ago (3 children)

I swear every single article critical of current LLMs is like, "The square got BLASTED by the triangle shape when it completely FAILED to go through the triangle shaped hole."

[–] drspod@lemmy.ml 42 points 1 day ago (4 children)

It's newsworthy when the sellers of squares are saying that nobody will ever need a triangle again, and the shape-sector of the stock market is hysterically pumping money into companies that make or use squares.

[–] inconel@lemmy.ca 19 points 1 day ago (1 children)

It's also from a company claiming they're getting closer to create morphing shape that can match any hole.

And yet the company offers no explanation for how, exactly, they're going to get wood to do that.

load more comments (3 replies)
load more comments (2 replies)
[–] FMT99@lemmy.world 287 points 2 days ago (24 children)

Did the author thinks ChatGPT is in fact an AGI? It's a chatbot. Why would it be good at chess? It's like saying an Atari 2600 running a dedicated chess program can beat Google Maps at chess.

[–] spankmonkey@lemmy.world 227 points 2 days ago (28 children)

AI including ChatGPT is being marketed as super awesome at everything, which is why that and similar AI is being forced into absolutely everything and being sold as a replacement for people.

Something marketed as AGI should be treated as AGI when proving it isn't AGI.

load more comments (28 replies)
[–] malwieder@feddit.org 27 points 1 day ago (5 children)

Google Maps doesn't pretend to be good at chess. ChatGPT does.

load more comments (5 replies)
[–] iAvicenna@lemmy.world 16 points 1 day ago (1 children)

well so much hype has been generated around chatgpt being close to AGI that now it makes sense to ask questions like "can chatgpt prove the Riemann hypothesis"

load more comments (1 replies)
[–] merdaverse@lemm.ee 5 points 1 day ago* (last edited 1 day ago) (1 children)

OpenAI has been talking about AGI for years, implying that they are getting closer to it with their products.

https://openai.com/index/planning-for-agi-and-beyond/

https://openai.com/index/elon-musk-wanted-an-openai-for-profit/

Not to even mention all the hype created by the techbros around it.

load more comments (1 replies)
load more comments (20 replies)
[–] MonkderVierte@lemmy.zip 41 points 1 day ago (1 children)

LLM are not built for logic.

[–] PushButton@lemmy.world 18 points 1 day ago (2 children)

And yet everybody is selling to write code.

The last time I checked, coding was requiring logic.

[–] jj4211@lemmy.world 10 points 1 day ago (4 children)

To be fair, a decent chunk of coding is stupid boilerplate/minutia that varies environment to environment, language to language, library to library.

So LLM can do some code completion, filling out a bunch of boilerplate that is blatantly obvious, generating the redundant text mandated by certain patterns, and keeping straight details between languages like "does this language want join as a method on a list with a string argument, or vice versa?"

Problem is this can be sometimes more annoying than it's worth, as miscompletions are annoying.

load more comments (4 replies)
load more comments (1 replies)
[–] Sidhean@lemmy.blahaj.zone 10 points 1 day ago

Can i fistfight ChatGPT next? I bet I could kick its ass, too :p

[–] AlecSadler@sh.itjust.works 61 points 2 days ago (12 children)

ChatGPT has been, hands down, the worst AI coding assistant I've ever used.

It regularly suggests code that doesn't compile or isn't even for the language.

It generally suggests AC of code that is just a copy of the lines I just wrote.

Sometimes it likes to suggest setting the same property like 5 times.

It is absolute garbage and I do not recommend it to anyone.

[–] j4yt33@feddit.org 17 points 2 days ago (5 children)

I find it really hit and miss. Easy, standard operations are fine but if you have an issue with code you wrote and ask it to fix it, you can forget it

load more comments (5 replies)
[–] arc99@lemmy.world 8 points 1 day ago (2 children)

All AIs are the same. They're just scraping content from GitHub, stackoverflow etc with a bunch of guardrails slapped on to spew out sentences that conform to their training data but there is no intelligence. They're super handy for basic code snippets but anyone using them anything remotely complex or nuanced will regret it.

load more comments (2 replies)
load more comments (10 replies)
[–] Objection@lemmy.ml 82 points 2 days ago (5 children)

Tbf, the article should probably mention the fact that machine learning programs designed to play chess blow everything else out of the water.

[–] andallthat@lemmy.world 13 points 1 day ago* (last edited 1 day ago)

Machine learning has existed for many years, now. The issue is with these funding-hungry new companies taking their LLMs, repackaging them as "AI" and attributing every ML win ever to "AI".

ML programs designed and trained specifically to identify tumors in medical imaging have become good diagnostic tools. But if you read in news that "AI helps cure cancer", it makes it sound like it was a lone researcher who spent a few minutes engineering the right prompt for Copilot.

Yes a specifically-designed and finely tuned ML program can now beat the best human chess player, but calling it "AI" and bundling it together with the latest Gemini or Claude iteration's "reasoning capabilities" is intentionally misleading. That's why articles like this one are needed. ML is a useful tool but far from the "super-human general intelligence" that is meant to replace half of human workers by the power of wishful prompting

[–] bier@feddit.nl 28 points 2 days ago (2 children)

Yeah its like judging how great a fish is at climbing a tree. But it does show that it's not real intelligence or reasoning

load more comments (2 replies)
[–] Zenith@lemm.ee 14 points 1 day ago

I forgot which airline it is but one of the onboard games in the back of a headrest TV was a game called “Beginners Chess” which was notoriously difficult to beat so it was tested against other chess engines and it ranked in like the top five most powerful chess engines ever

load more comments (2 replies)
[–] finitebanjo@lemmy.world 15 points 1 day ago

All these comments asking "why don't they just have chatgpt go and look up the correct answer".

That's not how it works, you buffoons, it trains off of datasets long before it releases. It doesn't think. It doesn't learn after release, it won't remember things you try to teach it.

Really lowering my faith in humanity when even the AI skeptics don't understand that it generates statistical representations of an answer based on answers given in the past.

[–] arc99@lemmy.world 20 points 1 day ago (1 children)

Hardly surprising. Llms aren't -thinking- they're just shitting out the next token for any given input of tokens.

[–] stevedice@sh.itjust.works 1 points 8 hours ago

That's exactly what thinking is, though.

[–] Furbag@lemmy.world 29 points 2 days ago (6 children)

Can ChatGPT actually play chess now? Last I checked, it couldn't remember more than 5 moves of history so it wouldn't be able to see the true board state and would make illegal moves, take it's own pieces, materialize pieces out of thin air, etc.

load more comments (6 replies)
[–] floofloof@lemmy.ca 44 points 2 days ago* (last edited 2 days ago) (5 children)

I suppose it's an interesting experiment, but it's not that surprising that a word prediction machine can't play chess.

load more comments (5 replies)
[–] anubis119@lemmy.world 36 points 2 days ago (5 children)

A strange game. How about a nice game of Global Thermonuclear War?

load more comments (5 replies)
load more comments
view more: next ›