Get live statistics and analysis of Alex Dimakis's profile on X / Twitter

Professor, UC berkeley | Founder @bespokelabsai |

2k following21k followers

The Analyst

Alex Dimakis is a sharp and detail-oriented academic and AI founder who thrives on dissecting complex AI behaviors with precision and curiosity. As a professor at UC Berkeley and founder of Bespoke Labs AI, Alex blends deep theoretical insights with practical innovation to push the boundaries of understanding in machine learning. His tweets reveal a passion for unraveling the nuances of AI models and a commitment to fostering realistic expectations around their capabilities.

Impressions
35.6k9.8k
$6.68
Likes
2584
74%
Retweets
27-6
8%
Replies
12-7
3%
Bookmarks
5213
15%

Top users who interacted with Alex Dimakis over the last 14 days

@cryptodaaddy

Growth Strategist • Web3 Investor | 9-Figure Vision • EX @ezu_xyz

1 interactions

Alex probably proofreads his own tweets with a spellchecker set to 'nitpicky professor mode'—so much detail that even robots get overwhelmed and start doubting their own reasoning skills.

Alex has successfully combined an academic career at a top institution with founding an AI startup, while maintaining influential thought leadership through detailed, high-engagement tweets dissecting the nuances of modern AI systems.

To deepen the collective understanding of AI’s capabilities and limitations, bridging theoretical research with real-world applications, while educating and challenging the AI community to think critically about model reasoning and performance.

Alex values scientific rigor, transparency, and intellectual honesty. He believes in thorough empirical analysis, embracing nuance over hype, and the importance of advancing AI responsibly through careful scrutiny. He is skeptical of oversimplifications and champions a nuanced, data-driven approach to AI research.

Alex’s greatest strength lies in his exceptional analytical mind and ability to communicate complex AI research insights clearly. His academic background combined with entrepreneurial experience allows him to critically assess AI models while influencing the field with fresh, practical ideas.

Tending toward deep technical dives and critical scrutiny, Alex sometimes risks coming across as overly cautious or skeptical, potentially limiting his appeal to audiences craving more optimistic or simplified AI narratives.

On X, Alex should leverage his expertise by sharing thread-style deep dives that break down complex AI topics with accessible analogies, paired with engaging visuals or simplified summaries. Regular interactive Q&A sessions could boost engagement and attract followers interested in thoughtful AI discourse.

Alex often highlights surprising weaknesses in state-of-the-art models, such as GPT-4’s struggles with basic multiplication and the counterintuitive observation that wrong reasoning model answers tend to be longer than correct ones.

Top tweets of Alex Dimakis

"RL with only one training example" and "Test-Time RL" are two recent papers that I found fascinating. In the "One Training example" paper the authors find one question and ask the model to solve it again and again. Every time, the model tries 8 times (the Group in GRPO), and a gradient step is performed, to increase the reward which is a very simple verification of the correct answers, repeated thousands of times on the same problem. The shocking finding is that the model does not overfit to this one question: RL on one example, makes the model better in MATH500 and other benchmarks. (If instead you did SFT repeating one training question-solution finetuning, the model would quickly memorize this answer and overfit). But with RL, the model has to solve the problem itself, since it only sees the question, not the answer. Every time it produces different answers, and this seems to prevent overfitting. The other papers are relying on the same phenomenon: you can have a small number of training questions and re-solve them thousands of times. You can do this for the test set (as test-time RL does) and still not overfit. We also independently saw this by doing RL training on half the test set and seeing benefits in the other half for BFCL agents. My thought now is that this shows our RL learning algorithm must be extremely inefficient. When a human is learning by solving a math puzzle, they immediately learn what they can learn by solving it once (or twice). No further benefit would come by assigning the same homework problem to students a tenth time. But in RL, we keep asking the model to re-solve the same question thousands of times, and the model slowly gets better. We should be able to have much better RL learning algorithms since the information is there. (1/2)

349k

youtube.com/watch?v=zjkBMF… Probably the best 1h introduction to LLMs that I've seen. And after 20mins its not an introduction, its getting into cutting edge research updates updated up to this month. I had not heard of the data exfiltration by prompt injection or the recent finetuning Poisoning attacks.

73k

Most engaged tweets of Alex Dimakis

"RL with only one training example" and "Test-Time RL" are two recent papers that I found fascinating. In the "One Training example" paper the authors find one question and ask the model to solve it again and again. Every time, the model tries 8 times (the Group in GRPO), and a gradient step is performed, to increase the reward which is a very simple verification of the correct answers, repeated thousands of times on the same problem. The shocking finding is that the model does not overfit to this one question: RL on one example, makes the model better in MATH500 and other benchmarks. (If instead you did SFT repeating one training question-solution finetuning, the model would quickly memorize this answer and overfit). But with RL, the model has to solve the problem itself, since it only sees the question, not the answer. Every time it produces different answers, and this seems to prevent overfitting. The other papers are relying on the same phenomenon: you can have a small number of training questions and re-solve them thousands of times. You can do this for the test set (as test-time RL does) and still not overfit. We also independently saw this by doing RL training on half the test set and seeing benefits in the other half for BFCL agents. My thought now is that this shows our RL learning algorithm must be extremely inefficient. When a human is learning by solving a math puzzle, they immediately learn what they can learn by solving it once (or twice). No further benefit would come by assigning the same homework problem to students a tenth time. But in RL, we keep asking the model to re-solve the same question thousands of times, and the model slowly gets better. We should be able to have much better RL learning algorithms since the information is there. (1/2)

349k

github.com/mlfoundations/… I’m excited to introduce Evalchemy 🧪, a unified platform for evaluating LLMs. If you want to evaluate an LLM, you may want to run popular benchmarks on your model, like MTBench, WildBench, RepoBench, IFEval, AlpacaEval etc as well as standard pre-training metrics like MMLU. This requires you to download and install more than 10 repos, each with different dependencies and issues. This is, as you might expect, an actual nightmare. (1/n)

146k

People with Analyst archetype

The Analyst

studying AI and trust. ex @openai/@googledeepmind

1k following65k followers
The Analyst

Research @allen_ai, reasoning, open models, RL(VR/HF)... Contact via email. Writes @interconnectsai, Wrote The RLHF Book, 🏔️🏃‍♂️

885 following60k followers
The Analyst

👾 Researcher 👾Chasing bubbles 👾 Back to my X arc 👾

2k following6k followers
The Analyst

Sap Consulting Development Agent supervision Vibecoding Vibe-Code cleaner ;)

122 following127 followers
The Analyst

Former Google & Slack M&A, Ice cream maker & gamer.

715 following2k followers
The Analyst

I Research Defi,Info fi & Web3 📚/ Ex @wasabi_protocol / Moon Cycle Trader 🌖

5k following6k followers
The Analyst

sentio ergo sum. science @METR_Evals

1k following3k followers
The Analyst

mathematical physicist lost in an extraliminal space of events to take humanity toward the singularity of knowledge 📙 aspiring timelord

1k following22k followers
The Analyst

Moving away from memes and into UTILITY only projects . DYOR Everything I post is my own opinion and not financial advice!!!

7k following3k followers
The Analyst

Investor 💵 | /MES Futures Trader 👨‍💻 | 5+Years of Stock Market Experience 📊 | Business Degree 👨‍🎓 | Dividend Tracker 💸 | ETF Nerd 🌱 | Spaces Host🎙

2k following8k followers
The Analyst

DLG_Crypto Ⓜ️Ⓜ️T

272 following209 followers
The Analyst

iresearch • iturn words into arts •TARSI • 🙂‍↔️🕊️ • itweet my thoughts and say what i think about protocols @RialoHQ @gensynai

1k following2k followers

Explore Related Archetypes

If you enjoy the analyst profiles, you might also like these personality types:

Supercharge your 𝕏 game,
Grow with SuperX!

Get Started for Free