Pragmatic Learning Theory, using tools from probability and Statistics | PhD in Stats @warwickstats 🇬🇧 | MMathStat @warwickstats 🇬🇧yuanhez.github.ioJoined September 2022
OpenAI hasn’t open-sourced a base model since GPT-2 in 2019. they recently released GPT-OSS, which is reasoning-only...
or is it?
turns out that underneath the surface, there is still a strong base model. so we extracted it.
introducing gpt-oss-20b-base 🧵
Another bad news for reasoning LLMs 🤔
The paper claims Chain-of-Thought in Language Models, is a brittle mirage bounded by training data, which is just pattern matching rather than genuine inference. 🤯
Argues that chain of thought in LLMs is pattern replay bound to training…
Another bad news for reasoning LLMs 🤔
The paper claims Chain-of-Thought in Language Models, is a brittle mirage bounded by training data, which is just pattern matching rather than genuine inference. 🤯
Argues that chain of thought in LLMs is pattern replay bound to training… https://t.co/t56iTm5p4d
Steal my ChatGPT prompt to master any topic using Feynman technique.
--------------------------------
FEYNMAN LEARNING COACH
--------------------------------
#CONTEXT:
Adopt the role of breakthrough learning architect. The user struggles with complex concepts that traditional…
"A Unified Theory of Language"
The paper argues language is a fast Bayesian pattern system, shaped by sexual selection to display intelligence.
It uses Construction Grammar, where a construction is a stored pairing that links form to meaning across words and gestures.…
Again another paper on the line of "Illusion" of thinking abilities of LLMs, this time from Japan. 😃
The author’s core logic is that true reasoning requires 100% guaranteed correctness, where premises must give conclusive relevant evidence for the conclusion, and LLMs can never…
A useful thing that GPT-5 can do that wasn’t previously possible before powerful AI is to monitor complex topics by asking it to give you scheduled reports.
Example: I have a weekly report on “reproducible, benchmarked evidence of autonomous or recursive self‑improvement in AI”
You don’t need GPT-5 or Claude 5...
You need better prompts.
MIT just confirmed what AI experts already knew:
Prompting drives 50% of performance.
Here’s how to level up without touching the model:
This one paper might kill the LLM agent hype.
NVIDIA just published a blueprint for agentic AI powered by Small Language Models.
And it makes a scary amount of sense.
Here’s the full breakdown:
Is Chain-of-Thought Reasoning of LLMs a Mirage?
... Our results reveal that CoT reasoning is a brittle mirage that vanishes when it is pushed beyond training distributions. This work offers a deeper understanding of why and when CoT reasoning fails, emphasizing the ongoing…
Beautiful @GoogleResearch paper.
LLMs can learn in context from examples in the prompt, can pick up new patterns while answering, yet their stored weights never change.
That behavior looks impossible if learning always means gradient descent.
The mechanisms through which this…
Confused about recent LLM RL results where models improve without any ground-truth signal? We were too. Until we looked at the reported numbers of the Pre-RL models and realized they were serverely underreported across papers. We compiled discrepancies in a blog below🧵👇
982 Followers 631 FollowingPh.D. student working of the foundations of AI/ML at @Penn.
Previously M.A. in Statistics @Wharton @Penn & B. Sc. in EE @ Sharif.
QR Intern @ Point72 (Cubist).
3K Followers 3K FollowingTweets about Bayesian statistics, Monte Carlo, etc from a Reader in Statistics at the University of Warwick. Personal account.
1K Followers 244 FollowingDepartment of Statistics, University of Warwick. Home of MORSE, CRiSM, AS&RU and APTS. Top 20 QS World University Rankings in Statistics & Operational Research.
3K Followers 744 FollowingDirector UCL Centre for AI and UiPath Distinguished Scientist. Co-founder https://t.co/Wx3VpUByR2. Pro: cycling, walking, EU. @[email protected]. Views my own.
268 Followers 230 FollowingAssistant Professor @WarwickDCS🇬🇧 | prev. @EPFL🇨🇭@KU_Leuven🇧🇪 @sjtu1896🇨🇳| math foundations of ML| co-organizer Foundation of AI seminar @FAIS_Warwick
163K Followers 166 FollowingCo-founder of Thinking Machines Lab @thinkymachines; Ex-VP, AI Safety & robotics, applied research @OpenAI; Author of Lil'Log
595K Followers 47K FollowingGallery of all things aesthetically pleasing 📸 images from multiple sources online | DM for credits, author claims or inquiries.
982 Followers 631 FollowingPh.D. student working of the foundations of AI/ML at @Penn.
Previously M.A. in Statistics @Wharton @Penn & B. Sc. in EE @ Sharif.
QR Intern @ Point72 (Cubist).
16K Followers 495 FollowingHarvard Professor.
Full stack ML and AI.
Co-director of the Kempner Institute for the Study of Artificial and Natural Intelligence.
3K Followers 388 FollowingI like Physics, Statistics, Machine learning, Computer Science & above all playing 🎸. Happy dad 👧 👧. Also professor @ EPFL. Views are my own.
691 Followers 664 Followingwork on foundations of AI, MLLM reliability/Eval, optimization, probability/stats, AI 4 math/science/med; Prof & director of center on AIF4S @USC 🚲🏔️🥾🏊♂️
18K Followers 4K FollowingAssociate Professor at UC Berkeley. Former Research Scientist at Google DeepMind. ML/AI Researcher working on foundations of LLMs and deep learning.
3K Followers 531 FollowingProfessor of CSC at @Concordia (CRC chair) & @Mila_Quebec. Visiting prof @AIatMeta. Previously @AIatMeta, @criteo, @inria. Interested in the principles of ML.