NireBryce

reality is the battlefield

the first line goes in Cohost embeds

๐Ÿฅ I am not embroiled in any legal battle
๐Ÿฆ other than battles that are legal ๐ŸŽฎ

I speak to the universe and it speaks back, in it's own way.

mastodon

email: contact at breadthcharge dot net

I live on the northeast coast of the US.

'non-functional programmer'. 'far left'.

conceptual midwife.

https://cohost.org/NireBryce/post/4929459-here-s-my-five-minut

If you can see the "show contact info" dropdown below, I follow you. If you want me to, ask and I'll think about it.


lexyeevee
@lexyeevee

i don't know what else to call it.

i ran across an excerpt from a paper recently, with a title mentioning something like "sparks of AGI" in GPT-4. the excerpt was about how they gave GPT-4 a modified problem from the 2022 IMO (a prestigious, and difficult, math contest) and it provided a solution and proof. holy fuck right

i looked at the proof and

  1. it was not a good proof. i don't want to get into it again but... it starts out seemingly okay, but then it goes kind of off the rails. and even the seemingly okay stuff seems dubious. the paper calls it a "correct proof" but i am highly skeptical.

    maybe i will Get Into It in another post idk. harder to do math on cohost with no mathjax

  2. i went to write my own solution for the problem and realized that the GPT proof had missed the whole point of the problem, the key insight about it. it really just handwaved and jumped ahead.

    and keep in mind: this was the answer they cherry-picked for their paper, and it was still junk. just plausible junk.

  3. to test my "plausible junk" hypothesis, i went and found the original problem and pasted it to GPT-4. if you are curious, the real problem is:

    Let โ„โบ denote the set of positive real numbers. Find all functions f : โ„โบ โ†’ โ„โบ such that for each x โˆˆ โ„โบ, there is exactly one y โˆˆ โ„โบ satisfying x f(y) + y f(x) โ‰ค 2.

    i don't have a fucking clue where to start with this. god damn. it is a thousand times more difficult than the baby high school problem they watered it down to.

    but GPT-4 gave me an answer, and wrote a proof for it! it was even formatted beautifully with latex!

    just one problem: the answer was completely wrong, and the proof contains basic algebra mistakes! it even contains "we will now prove that this solution is unique ... therefore, it is unique" without even attempting to prove that

and what really gets under my skin here is that, if they watered down this problem to make a version GPT-4 could solve, they must have tried the original first. surely, right? like if you're studying whether a computer can solve math problems, i can't imagine not even wondering how it would do with the real problem. that was the first thing that came to mind for me!

so why was that not in the fucking paper? what kind of "paper" only contains specific toy examples that you think worked well?

and the answer is a "paper" from microsoft, the same company that has poured ten billion dollars into openai so they can make stuff like gpt-4.

but they don't want to show the attempt to solve the original problem, because it demonstrates that what gpt-4 is actually designed to do is generate gibberish in the shape of human prose, and sometimes that prose coincidentally expresses a coherent idea. it's easier to steal the prestige of the IMO and staple that onto a high school calculus problem.

this whole field is a joke. it is hucksters and fraudsters and marketers. it is a fucking circus. it is people "studying" the thing while they have a strong financial interest in convincing the world at large that it's a thinking machine. it is advertising hosted on arxiv dot org. it is a scam and everyone involved in perpetuating it should be fucking ashamed.


the other paper to have crossed my path recently is this post about othello and whether a language model, trained only on sequences of moves, has an internal representation of the board or not.

it opens with a metaphor about a crow, because again, these people are trying to manipulate you into thinking that the computer has a brain now. it does not. it is a pattern-detection engine.

the experiment was to do the following:

  1. train a model on legal games of othello, represented as sequences of board coordinates such as F5 D4 A3 etc (i made that up it's probably not a legal game) edit: my bad they actually used completely arbitrary tokens so it couldn't infer the arrangement of the cells from their names
  2. train 64 more models on the original model, one per cell of the board, to see if they can detect anything that looks like a notion of what is in the specific cell
  3. compare this to a crow and conclude that the model has learned about the board state. give us funding

i'm not qualified to judge the gritty details, especially based on a funny png they made of some nodes, but my takeaway so far is this:

othello is a really mathematical game. i mean you've got a square grid and pieces that toggle in parity. it seems rife for patterns, especially if you throw a pattern-detection engine at it.

for example: every legal move has to be next to an existing piece. well, that's easy; we can find a bunch of legal moves with no clue what the board looks like, just by looking at the moves that have already been made! if D3 was a move, then the cells next to it are C2, C3, C4, D2, D4, E2, E3, and E4. but you can't play in the same cell twice, so eliminate any that are already in the move list. and there are rules about playing next to opposing colors, so if D3 was an even number of moves ago, then it's probably out. we've already culled huge numbers of illegal moves here โ€” there are surely other kinds of constraints to be inferred from the rules that would let you make legal moves based only on a transcript of the game. and you could do it without even knowing what the rules are, if you were some kind of pattern-detection engine being fed a lot of transcripts.

remember, this wasn't about whether it could play the game well, only whether it could make a move at all.

and you know that's kind of neat. it would be even neater if it were compared to any kind of previous mathematical analysis of othello, which almost certainly exists, but which this blog post probably doesn't even attempt to mention because it would make othello sound like a math thing which can be analyzed rather than a human thing which we've taught a computer to do.

and even the second model that can find some notion of grid cells in the first model is kind of neat.

but nobody can just stop there at "look what i made the thing do". no we have to claim we've found proof of an inner world. so now i have to spell out what actually happened here:

  1. they trained a model on legal games of othello, formatted as grid coordinates
  2. they trained a second model โ€” also a pattern-finding engine โ€” to specifically look for a notion of grid positions in the first model, based on what they knew the board should look like
  3. it worked

doesn't this seem a little bit like telling the horse when to stop counting? if you already know what shape of thing you're looking for, and you ask a model to find it... that's... what they do. it doesn't mean the first model contains a direct representation of the board; it means the first model contains something from which you can derive a representation of the board. and, fucking, of course it does โ€” because you can derive the board from the list of moves, which the first model contains! all you did was teach the second model how to ferret that out, based on the board state that you already knew!

am i missing something here because i swear to god!!!


NireBryce
@NireBryce

which is that those specific researchers at microsoft have never touched another field, and just accepted authoritative language.

Which I think might be the worse option.

But also: I'm surprised by how many people fall for things where it is clearly pulling the answers from blogs writing about the topic of your prompt, and often being wrong, as blogs are wont to do


You must log in to comment.

in reply to @lexyeevee's post:

my fucking NES could solve Othello to a degree sufficient to beat my ass on the regular. Hell I had a fucking TI calculator that could.

and neither of those things burned down a small forest in Ecuador to do it.

This shit should be fucking illegal.

the math paper is dumb but the Othello thing seems legitimately interesting to me. if you can literally adjust the model's perception of the world by modifying regions in its internal memory, that suggests pretty strongly that it is "storing the state". sure, you have to use multi-layer perceptrons for the interpretation, but it's not like when a human looks at an Othello board there's a nice 8x8 neuron grid.

ultimately I think it boils down to the fact that "know" and "understand" aren't really semantically meaningful concepts at a very low level. if I have two programs, one of which is a hand-coded script that I wrote that explicitly encodes the board state as an array of arrays and one of which is a big machine learning thing, and both of them always make valid moves, what does it mean to say one "understands" the rules and the other doesn't?

the model doesn't always make valid moves. it mostly makes valid moves. so it has some notion of rules that usually overlap with the actual rules. and no matter how much you train it you can never prove that it will always make valid moves.

but your script doesn't understand the rules either. it does encode them, though. the model doesn't.

i don't get the impulse to suggest that "understanding" means nothing. why? because we can poorly fake it with a massive enough database? do properties and relationships and mechanisms mean nothing? isn't it a pretty dark ages kind of thing to just write down stuff that happens in a big list and guess that it'll happen again similarly, no further insight required?

  • if i make an othello move, i can justify why it's legal (or not). the model can only suggest it was statistically likely to be a move (or not).
  • i can play othello without having ever seen a single game play out before.
  • i can grasp that a move is valid according to the rules even if i've never seen a similar move made before

anyway i think calling this "storing the board state" is making an anthropomorphizing leap โ€” because of course if you encode the board state then it must be in some sensible form. what else would you do? the blog post even makes a very strong direct comparison right at this point to its metaphor of a crow emulating the board in a physical grid.

but the model doesn't even know there's a grid. from a further read of the paper, what they demonstrated is that there's a linear function from some stuff in the model to the state of a particular cell.

and clearly the author really really wants to interpret this as "it understands the rules, it's simulating the game" but even the paper does ZERO discussion or citation of othello theory or anything to actually establish this

like is there a linear function from the set of currently-legal moves to the state of the board? i don't know. that seems kind of important to establish. surely mathematicians have looked into stuff like this. but it seems like all the citations are about other machine learning stuff.

which i guess is part of my recurring problem here โ€” it's like this paper is only interested in hinting at some philosophical implication about the model. it's not interested in the problem it used as fodder. as if everything that exists in the outside world were merely a homogeneous slurry of Training Data with no further context or history or people knowing things about it

hell it's not even interested in explaining what's actually going on. what does the model contain, exactly? what precisely is it gleaning? what's different about the tiny fraction of cases it fails on? well who cares about any of that; the point is, it's like a crow

maybe if the authors tried to understand, not just know

humans make invalid moves when playing games; chatgpt's typo rate is much much lower than mine, and my computer beats the hell out of me at adding 512-bit numbers.

anyway, it seems deeply unlikely to me that there's a linear function from board state to legal moves, but that doesn't seem relevant; presumably theres some horrendous function from my own neuron activity that recovers the state of an Othello board i was looking at, but I wouldn't expect that to look anything like the function you'd come up with if you just treated the board state as an array.

at no point does the post or the underlying paper claim that the storage format is human-comprehensible. but considering that you can use the probes to adjust what the model "thinks" the board state is in a way that causes it to adjust moves in line with the new board state, and that the new moves are sensible even if the board state is illegal, i think there is something going on. it's just difficult to figure out how these things work because interpretability of large neural networks is still a hard problem, and it seems like you're expecting them to completely solve it rather than go "hey, we took an existing technique and extended it to this new domain". in particular I actually used to work with two of the authors and they've done a lot of data visualization work in general.

also I just went and reread the blog post and they explicitly say that they aren't saying that the crow (in the analogy) or the model "understands" the rules, just that it has a world-state that is interpretable and controllable; the paper also only uses the word to refer to humans understanding things. so that's a digression entirely, oops

the actual thing the post claims to want to demonstrate is seemingly very mundane: that the model is storing something stateful, not just basically doing a markov chain on moves. but that seems kind of straightforward to me given the fact that it can generate legal moves remotely reliably at all, since othello is highly sensitive to ordering

and so what seems most interesting to me about this is that it is possible to essentially do a bigass statistical regression on sequences of moves and be able to extract individual cells after the fact. hey cool. honestly that seems to say a more interesting thing about the game of othello than about the math used to find it out

and maybe i've just become especially exhausted here, but spending four paragraphs describing a crow who makes their own little copy of the board sure seems like it is trying to imply both more about the internal model and some kind of philosophical thing about machine learning. the post even contains the line:

Finding these probes is like discovering the board made of seeds on the crow's windowsill.

what is a layperson going to take away from this? "oh it learned how to play the game". but no it didn't. the article paints a picture of a crow creating a physical grid but the model doesn't even have any concept of there being a grid at all. why does this kind of framing even come up? why can we not talk about what actually happened? it learned to generate legal moves, and from the thicket of stuff it tracks to do that, we can readily determine the current board state

by way of demonstration, this paper specifically found its way to me when someone linked it in a channel along with comments like:

(experiment that provides evidence that LLMs can construct a spatial model from text)

spatial! a spatial model! that's what happens when you draw a tight comparison to an animal, known to be intelligent, constructing a physical board.

and there's just so much of this. everyone wants to jump to the "it's alive!" analogies and no one wants to tone it down and it feels so willfully deceptive

edit: also i would think humans err at othello because they've read the board incorrectly (human eyesight is not ideal for scanning diagonally along a grid), not because they've misunderstood the rules. that's supposed to be the kind of thing computers are great at

It bothers me that I'm seeing a lot of smart-adjacent people I know being gulled by this hype and if I didn't work in software maybe I would be too.

Unlike the similar NFT hype last year, there is a baby in this bathwater, but it is definitely not the baby they are trying to sell us.

god reading that attempted proof it is so obvious that the model is just regurgitating patterns it has seen before. of course it says to plug in x = y because that's a common strategy in "find all functions satisfying this condition" problems but in this case it requires assuming that P(x, x) holds in the first place and so it's useless in further analysis

yeah and i think one of the big takeaways from all this is that "throw everything at the wall and see what sticks" will actually get you pretty far, but language models just keep throwing since they have no notion of whether anything's actually sticking

thereโ€™s a lot of fraud that you can just, get away with for free, if the people youโ€™re swindling donโ€™t bother to do even basic research into validity of claims and conflicts of interest, and itโ€™s very frustrating to watch from the sidelines. idk, people in comp tech industries think theyโ€™re super smart, and yet swaths of them fall for this shit. it makes me think of the whole Wata Games scam, which is perpetuated by the same types of people who sell their own comic books to themselves periodically to make it appear to be gradually appreciating in value (to anyone who doesnโ€™t bother to check whoโ€™s on both sides of these transactions). other things that come to mind are anything that claims to be part of โ€œweb3โ€, the various antiโ€science movements, and i suppose Tommy Tallarico in his entirety. itโ€™s unfortunate to see machine learning join these sorry bunches, because as a technology it has genuinely useful applications when utilized properly. but those uses are being readily ignored by big comp tech because identifying malignant tumors isnโ€™t nearly as flashy and investorโ€wowing as uh, chatbots? i guess?

the thing that really gets me about all this is the way that, like the whole nft boom, the field is full of people going "behold this brilliant achievement!" and then you point out hey, this looks/reads/calculates like rubbish, and they go "no it's good", and that's the matter apparently settled. it's genuinely unsettling to me how many DALL-E prompts produce images that are simply, objectively, not what was prompted, despite everyone acting like they are. do people genuinely just... not notice that they're not getting the meal they ordered?

This is why Dall-E generates 9 or 10 images at a time. The miss rate is only about 50%, so you wind up with something like an 85-90% probability that they did get the meal they ordered... and all the other shit on the side isn't mixed in, so they can throw it away without noticing just how much of it there is!

I don't really think that's true though, I think the miss rate is much closer to 90% or higher and people are just gaslighting themselves into seeing something that simply is not there. like I saw someone plug in the prompt "american gothic, with two dogs holding pepperoni pizza instead of the farmers holding the pitchforks." every single result was two dogs dressed as farmers holding pitchforks with a pepperoni pizza nearby. occasionally one was holding it. none of the pieces looked much at all like American Gothic the famous painting. to me, particularly as someone who can draw, that feels like a 100% miss rate!

maybe you're onto something though that just presenting 10 images at once means that instead of seeing the images as individual failures, people are being tricked essentially into comparing them only to each other rather than to the ideal hypothetical image that would in fact fulfill the prompt. even if the best of the bunch is still, technically, not fulfilling the prompt, it sort of has the illusion of doing so because you're getting distracted by the other, worse images right in front of you that it is superior to.

As a maybe-interesting point of context, I've watched AI-focused companies insist that they were this close to "real" artificial intelligence for almost thirty years. Back then, it was companies with (mildly) fancy Prolog interpreters, sure that they only lacked a few facts, something that "everybody knows" that they've forgotten to load into their databases.

And everybody ignored them, because that's inherently laughable. Like, philosophers of every stripe have been studying how to model and replicate intelligence for thousands of years, and every one of these dolts thinks that it's going to happen for them accidentally!

On top of that, not one of these people wants to have a (legitimate and probably interesting) Chinese Room Problem discussion. And I think that's why these inane announcements actually feel so offensive: They're skipping over the "how do we define intelligence" part (which, let's be honest, has ramifications for animal rights and their rights over their creations) and insisting that they've found it (which gets them investment money).

I've long been of the opinion that the Chinese Room requires either actual understanding or a physically impossible infinitely large flowchart. And I feel like that's what we're seeing with GPT models. The flowchart can get larger and larger, and the percentage of "this isn't actually a conversation" style of errors gets smaller and smaller, but the category still remains, and remains significant.

i think an overlooked problem with most thought experiments here is that in practice, if you tell a human "this computer can talk", it turns out they'll have surface-level conversations about either nothing, basic factoids, or common quips: "hello", "how big is the sun", "this sentence is false" โ€” exactly the sort of stuff a model trained on the internet would have seen the most

and this is because you can't just sit a human in front of a computer and ask them to have another conversation. you can barely ask two humans to just have a conversation. it's practically a trope that speed or blind dating are hard, and this is why: what do you talk about? there's no shared context. people talk about things they have in common, or at least things they have reason to think each other are interested in. with a stranger, well, at least you can talk about the weather. with a computer, you don't even have that.

That's the interesting part, though, at least to me. It requires "understanding," but can the understanding be encoded into the physical form? Probably (we're finite creatures, after all), but probably not using existing math and 1960s technologies.

I mean, it doesn't matter how many layers of neural networks they pile on. It still needs to run on the same Turing-complete computer with the same limits to computation that it always had...

(Which is a long way of agreeing with you.)

It does really seem like the AI industry is playing a game of "Texas sharpshooter". How convenient that everyone seems to be close to "real" artificial intelligence, but no one has properly defined the target...

That said, I have mixed feelings about trying to define it. It still seems like the closest thing we have is the Turing test (which emphatically is only concerned with "imitation" intelligence). Arguably, AI has reached that goal, but the results feel cursed.

"I wish for a computer program that can imitate human communication"

"Wish granted... but it's because most human communication is now commodified"

Maybe that's why I like Dijkstra's answer so much. "The question of whether a computer can think is no more interesting than the question of whether a submarine can swim." Intelligence doesn't matter without practical situations to apply it to.

But even (or maybe especially) from that perspective, recent AI results feel very underwhelming. Theorem provers that don't actually prove theorems, othello bots that don't even get the rules right... these are not useful things. They're submarines that sink.

Right, and for clarity, I don't advocate a measurement of intelligence. We've done that with things like IQ tests, and it's always biased and excludes actual humans who we consider intelligent, in favor of people who...just study. I raised the issue of a definition, because (like you say) claiming to have detected something without a definition to measure it against is mostly just wishful thinking. They might as well claim that the AI is haunted.

And absolutely, while I find the AIs entertaining, especially as springboards to better ideas, I have yet to see them do anything that we didn't have in the 1990s, either with smaller neural networks or classic AI algorithms like Markov processes.

Welcome to academia! There's a lot of real bullshit papers out there that get snuck through peer review with grandiose language and overworked researchers. Depending on the subject and profession, can have between 0.08% to 40% to 70% of papers being unable to be replicated. https://en.wikipedia.org/wiki/Replication_crisis

Def feels like AI research due to it being in the public spotlight again has a lot of reasons to be very sneaky with language and make things seem bigger than they are. Easier to get grant and research funding with a bit of lying and omissions!

My favorite thing about the Sparks paper is that it's a non-peer reviewed paper with a primary author who is "Sr. Principal Research Manager in the Machine Learning Foundations group at Microsoft Research (MSR)". It's ad copy! Written by a manager!

I love machine learning and I think it's amazing... but yeah, this is interesting. I get a lot of anti-AI stuff on my feed and I usually ignore it because... I don't need another negative post with a lot of bias about why AI sucks. It usually doesn't tell me (those not in the loop) why machine learning has a lot of fraud.

This does. I pondered my own biases, and thought about what models really are. It made sense and I quite liked that it made sense. Thank you for explaining these papers, but also showing some positives.

i don't actually understand what the point of using a mapping to arbitrary tokens for Othello moves would be, given that these models receive tokens in a categorical form. whatever those categorical values map to is irrelevant, they might as well be legible to the people inputting the data. if this is a normal token LLM then the idea that the model could infer position based on the names of the tokens, which it cannot see or experience, is pure superstition that belies either incompetence and lack of understanding of the core of their paper, or willful misleading of the reader.

ok i looked at it and they are just not providing the board state to the model in any spatially reasoned way. i think alpha zero had convolution layers and so on that gave it a natural perception of the uniformity of the game space, and they just left that out here.

anyway you're right these papers are fuckin worthless lol. god i hate my industry