Tarun Chitra
@pinged
Part IV: Reasoning without Regret ~~~ Q: Can we quantify when we can make these models better: higher accuracy + lower compute cost? A: EIP-1559 is everywhere 😈 DeepSeek is a phase transition: Lowered compute by 10x+ with ~same accuracy as o1 — why? Must be real math & algorithmic improvement 🤓 🔫'd me
1 reply
0 recast
67 reactions
Tarun Chitra
@pinged
First off, I wouldn't have been able to solve this without having o3-mini; I used it to find references + ideas that I had never heard of — would have likely taken me forever to find on my own But there is something tantalizing about the idea of using a reasoning model to solve a math problem about a reasoning model itself; if you can do this, you found the 'backdoor' to the Reasoning Russell's paradox (in so far as one can convince themselves that the reasoning model can prove *some* properties about the set of possible reasoning traces it generates, even though it might not be able to describe the whole set) This became my rallying cry as a way to get out of the AI doomer 🕳️ — figure out what makes DeepSeek tick using o3 as an assistant (ironic, I know)
1 reply
0 recast
2 reactions
xuning
@xuning
A+
0 reply
0 recast
0 reaction