[-] kromem@lemmy.world 1 points 1 day ago

Actually, they are hiding the full CoT sequence outside of the demos.

What you are seeing there is a summary, but because the actual process is hidden it's not possible to see what actually transpired.

People are very not happy about this aspect of the situation.

It also means that model context (which in research has been shown to be much more influential than previously thought) is now in part hidden with exclusive access and control by OAI.

There's a lot of things to be focused on in that image, and "hur dur the stochastic model can't count letters in this cherry picked example" is the least among them.

[-] kromem@lemmy.world 20 points 3 days ago

I was thinking the same thing!!

It's like at this point Trump is watching the show to take notes and stage direction.

[-] kromem@lemmy.world 7 points 3 days ago* (last edited 3 days ago)

Yep:

https://openai.com/index/learning-to-reason-with-llms/

First interactive section. Make sure to click "show chain of thought."

The cipher one is particularly interesting, as it's intentionally difficult for the model.

The tokenizer is famously bad at two letter counts, which is why previous models can't count the number of rs in strawberry.

So the cipher depends on two letter pairs, and you can see how it screws up the tokenization around the xx at the end of the last word, and gradually corrects course.

Will help clarify how it's going about solving something like the example I posted earlier behind the scenes.

[-] kromem@lemmy.world 5 points 3 days ago

You should really look at the full CoT traces on the demos.

I think you think you know more than you actually know.

[-] kromem@lemmy.world -3 points 3 days ago* (last edited 3 days ago)

I'd recommend everyone saying "it can't understand anything and can't think" to look at this example:

https://x.com/flowersslop/status/1834349905692824017

Try to solve it after seeing only the first image before you open the second and see o1's response.

Let me know if you got it before seeing the actual answer.

[-] kromem@lemmy.world 69 points 4 days ago* (last edited 4 days ago)

I fondly remember reading a comment in /r/conspiracy on a post claiming a geologic seismic weapon brought down the towers.

It just tore into the claims, citing all the reasons this was preposterous bordering on batshit crazy.

And then it said "and your theory doesn't address the thermite residue" going on to reiterate their wild theory.

Was very much a "don't name your gods" moment that summed up the sub - a lot of people in agreement that the truth was out there, but bitterly divided as to what it might actually be.

As long as they only focused on generic memes of "do your own research" and "you aren't being told the truth" they were all on the same page. But as soon as they started naming their own truths, it was every theorist for themselves.

[-] kromem@lemmy.world 12 points 5 days ago* (last edited 5 days ago)

The pause was long enough she was able to say all the things in it mentally.

[-] kromem@lemmy.world 3 points 5 days ago

They got off to a great start with the PS5, but as their lead grew over their only real direct competitor, they became a good example of the problems with monopolies all over again.

This is straight up back to PS3 launch all over again, as if they learned nothing.

Right on the tail end of a horribly mismanaged PSVR 2 launch.

We still barely have any current gen only games, and a $700 price point is insane for such a small library to actually make use of it.

[-] kromem@lemmy.world 9 points 1 week ago* (last edited 1 week ago)

Meanwhile, here's an excerpt of a response from Claude Opus on me tasking it to evaluate intertextuality between the Gospel of Matthew and Thomas from the perspective of entropy reduction with redactional efforts due to human difficulty at randomness (this doesn't exist in scholarship outside of a single Reddit comment I made years ago in /r/AcademicBiblical lacking specific details) on page 300 of a chat about completely different topics:

Yeah, sure, humans would be so much better at this level of analysis within around 30 seconds. (It's also worth noting that Claude 3 Opus doesn't have the full context of the Gospel of Thomas accessible to it, so it needs to try to reason through entropic differences primarily based on records relating to intertextual overlaps that have been widely discussed in consensus literature and are thus accessible).

[-] kromem@lemmy.world 3 points 1 week ago

This is pretty much every study right now as things accelerate. Even just six months can be a dramatic difference in capabilities.

For example, Meta's 3-405B has one of the leading situational awarenesses of current models, but isn't present at all to the same degree in 2-70B or even 3-70B.

[-] kromem@lemmy.world 9 points 2 weeks ago

Lucretius in De Rerum Natura in 50 BCE seemed to have a few that were just a bit ahead of everyone else, owed to the Greek philosopher Epicurus.

Survival of the fittest (book 5):

"In the beginning, there were many freaks. Earth undertook Experiments - bizarrely put together, weird of look Hermaphrodites, partaking of both sexes, but neither; some Bereft of feet, or orphaned of their hands, and others dumb, Being devoid of mouth; and others yet, with no eyes, blind. Some had their limbs stuck to the body, tightly in a bind, And couldn't do anything, or move, and so could not evade Harm, or forage for bare necessities. And the Earth made Other kinds of monsters too, but in vain, since with each, Nature frowned upon their growth; they were not able to reach The flowering of adulthood, nor find food on which to feed, Nor be joined in the act of Venus.

For all creatures need Many different things, we realize, to multiply And to forge out the links of generations: a supply Of food, first, and a means for the engendering seed to flow Throughout the body and out of the lax limbs; and also so The female and the male can mate, a means they can employ In order to impart and to receive their mutual joy.

Then, many kinds of creatures must have vanished with no trace Because they could not reproduce or hammer out their race. For any beast you look upon that drinks life-giving air, Has either wits, or bravery, or fleetness of foot to spare, Ensuring its survival from its genesis to now."

Trait inheritance from both parents that could skip generations (book 4):

"Sometimes children take after their grandparents instead, Or great-grandparents, bringing back the features of the dead. This is since parents carry elemental seeds inside – Many and various, mingled many ways – their bodies hide Seeds that are handed, parent to child, all down the family tree. Venus draws features from these out of her shifting lottery – Bringing back an ancestor’s look or voice or hair. Indeed These characteristics are just as much the result of certain seed As are our faces, limbs and bodies. Females can arise From the paternal seed, just as the male offspring, likewise, Can be created from the mother’s flesh. For to comprise A child requires a doubled seed – from father and from mother. And if the child resembles one more closely than the other, That parent gave the greater share – which you can plainly see Whichever gender – male or female – that the child may be."

Objects of different weights will fall at the same rate in a vacuum (book 2):

“Whatever falls through water or thin air, the rate Of speed at which it falls must be related to its weight, Because the substance of water and the nature of thin air Do not resist all objects equally, but give way faster To heavier objects, overcome, while on the other hand Empty void cannot at any part or time withstand Any object, but it must continually heed Its nature and give way, so all things fall at equal speed, Even though of differing weights, through the still void.”

Often I see people dismiss the things the Epicureans got right with an appeal to their lack of the scientific method, which has always seemed a bit backwards to me. In hindsight, they nailed so many huge topics that didn't end up emerging again for millennia that it was surely not mere chance, and the fact that they successfully hit so many nails on the head without the hammer we use today indicates (at least to me) that there's value to looking closer at their methodology.

133
submitted 3 months ago by kromem@lemmy.world to c/technology@lemmy.world

I often see a lot of people with outdated understanding of modern LLMs.

This is probably the best interpretability research to date, by the leading interpretability research team.

It's worth a read if you want a peek behind the curtain on modern models.

9
submitted 5 months ago by kromem@lemmy.world to c/technology@lemmy.world
79
submitted 5 months ago by kromem@lemmy.world to c/technology@lemmy.world
7
submitted 7 months ago* (last edited 7 months ago) by kromem@lemmy.world to c/technology@lemmy.world

I've been saying this for about a year since seeing the Othello GPT research, but it's nice to see more minds changing as the research builds up.

Edit: Because people aren't actually reading and just commenting based on the headline, a relevant part of the article:

New research may have intimations of an answer. A theory developed by Sanjeev Arora of Princeton University and Anirudh Goyal, a research scientist at Google DeepMind, suggests that the largest of today’s LLMs are not stochastic parrots. The authors argue that as these models get bigger and are trained on more data, they improve on individual language-related abilities and also develop new ones by combining skills in a manner that hints at understanding — combinations that were unlikely to exist in the training data.

This theoretical approach, which provides a mathematically provable argument for how and why an LLM can develop so many abilities, has convinced experts like Hinton, and others. And when Arora and his team tested some of its predictions, they found that these models behaved almost exactly as expected. From all accounts, they’ve made a strong case that the largest LLMs are not just parroting what they’ve seen before.

“[They] cannot be just mimicking what has been seen in the training data,” said Sébastien Bubeck, a mathematician and computer scientist at Microsoft Research who was not part of the work. “That’s the basic insight.”

5
submitted 7 months ago by kromem@lemmy.world to c/chatgpt@lemmy.world

I've been saying this for about a year, since seeing the Othello GPT research, but it's great to see more minds changing on the subject.

204
submitted 10 months ago by kromem@lemmy.world to c/world@lemmy.world
9
submitted 1 year ago by kromem@lemmy.world to c/chatgpt@lemmy.world

I've suspected for a few years now that optoelectronics is where this is all headed. It's exciting to watch as important foundations are set on that path, and this was one of them.

2
submitted 1 year ago by kromem@lemmy.world to c/history@lemmy.world

The Minoan style headbands from Egypt during the 18th dynasty is particularly interesting.

view more: next ›

kromem

joined 1 year ago