pull down to refresh
100 sats \ 1 reply \ @Scoresby OP 23h \ parent \ on: I guess I was wrong about AI persuasion - Dynomight AI
Isn't this a bit like saying measuring all the various wavelengths bouncing off a painting can help you know whether or not it is art?
I think the question of AI reasoning may remain unanswerable because whatever is going on in an LLM is different enough from what goes on in a human mind that we'll never be at peace with the comparison.
What is my evidence for this? Probably something like Stanislaw Lem's Solaris -- which I admit is a little shaky. But, I don't think we have a good handle on what reasoning is in human minds, and so measurement or not, it seems to me that we won't be able to make a clear determination about LLM reasoning. It may be that some people think it counts as thinking and maybe others don't.
Isn't this a bit like saying measuring all the various wavelengths bouncing off a painting can help you know whether or not it is art?
Nice analogy!
I'd say to the philosopher it definitely is. It says in the article:
The “principled reasoner” being compared to here simply does not exist. It’s a Platonic ideal.
I agree with that. But what it feels like is that the "template" that is being reinforced during reasoning training 1, is maybe good for math and coding, but not so awesome for other domains. I think what's needed is new alignment policies that aren't solely derived from the experts that are OpenAI/Goog/Anthropic hires - some diversity in alignment would be nice! 2.
I don't think we have a good handle on what reasoning is in human minds
I think the problem is trying to emulate it in the first place. I'm personally always amazed by people that tick differently, and they help me break through barriers in my thinking. I don't need people that think like me around me. Compatible, but not similar?
Footnotes
-
I cannot help but feel that the method and scoring is the same or at least very similar for all LLMs minus Claude and Gemini at the moment, also because when DeepMind decided to do it wholly differently, it turned out that the results kind of sucked. ↩
-
I was fantasizing the other day that I'd totally want a mixture of experts LLM where each expert layer has been aligned by a different Bitcoiner... So you basically get the luke-jr layer arguing with the petertodd layer in the reasoning phase. That would be so awesome for those of us that quit the bird app... once a month, on demand. "you're wrong and you know it." -> "you lie!". hah! ↩
reply