- The Vision, Debugged;
- Posts
- AdaptThink: Teaching AI When NOT to Think š§
AdaptThink: Teaching AI When NOT to Think š§
PLUS: Why every video online could now be completely fake

Howdy Vision Debuggers! šµ
While most of us overthink our morning coffee choices, Spark and Trouble discovered something fascinating about AI that knows exactly when to hit the mental brakes and when to floor it.
Hereās a sneak peek into todayās edition š
Reasoning models finally learn when NOT to overthink problems
Todayās prompt might be your blueprint for viral content creation
3 powerful AI tools that will blow your mind
Googleās Veo3 just broke the internet with impossibly realistic fake videos
Time to jump in!š
PS: Got thoughts on our content? Share 'em through a quick survey at the end of every edition. It helps us see how our product labs, insights & resources are landing, so we can make them even better.

Hot off the Wires š„
We're eavesdropping on the smartest minds in research. 𤫠Don't miss out on what they're cooking up! In this section, we dissect some of the juiciest tech research that holds the key to what's next in tech.ā”
Remember those math tests in school where you'd encounter a mix of problemsāsome so simple you could answer instantly, while others required you to show all your work, step by step? You probably learned when to trust your gut and when to slow down and think it through.
Well, researchers at top AI labs have been wrestling with a similar challenge: teaching AI models (the recent flurry of powerful āreasoningā models like OpenAIās o-series, & DeepSeekās R-series) when they actually need to "think" and when they can just cut to the chase.
Enter AdaptThink, a breakthrough from researchers at Tsinghua University, that's solving one of the most practical problems in modern AI: the unnecessary overhead of overthinking simple problems. This isn't just another research paperāit's a game-changer for making AI reasoning both smarter and more efficient.
So, whatās new?
Today's reasoning models like GPT-o1 and DeepSeek-R1 are incredibly powerful, capable of working through complex problems with detailed chain-of-thought reasoning. But here's the catch: they're doing this elaborate thinking process for everything, even when asked, "What's 2+2?" It's like using a sledgehammer to crack a nutātechnically effective, but wildly inefficient.
This creates a real problem. Every token generated costs money and time. When a model generates a 500-token reasoning trace to solve a simple arithmetic problem that could be answered in three characters, you're looking at roughly 167x more computational cost than necessary. Scale this across millions of queries daily, and you've got a serious efficiency bottleneck.

AdaptThink tried to think only when needed, to achieve a balance between accuracy & efficiency (source: AdaptThink paper)
AdaptThink tackles this head-on by teaching models a crucial skill: knowing when to think and when to just answer. It's like training an AI to have the same intuitive judgment you developed during those school math testsārecognising when a problem is simple enough to solve directly versus when it requires careful, step-by-step reasoning.
Under the hoodā¦
AdaptThink's brilliance lies in its elegant simplicity. The system trains models to choose between two distinct modes:
Thinking Mode: The full chain-of-thought reasoning we're familiar with, complete with exploration, reflection, and step-by-step logicāperfect for complex problems that genuinely require deep analysis.
NoThinking Mode: Direct answer generation without the internal reasoning traceāideal for problems where the model is confident and the solution is straightforward.
The magic happens through a sophisticated reinforcement learning approach that uses constrained optimisation. Think of it as teaching the model to be efficient without sacrificing accuracy. The algorithm maximises the use of NoThinking mode while maintaining a hard constraint: performance must never drop below the original model's baseline.
Here's where it gets technically interesting. The researchers faced a classic chicken-and-egg problem: early in training, models almost never choose NoThinking mode because they haven't learned when it's appropriate.
To solve this, they implemented importance samplingāa technique that forces the model to explore NoThinking responses for 50% of training examples, preventing the dreaded mode collapse where the model gets stuck in always-thinking patterns.
The training objective is particularly clever. It rewards the model not just for getting answers right, but specifically for using NoThinking when it's safe to do so. The algorithm balances this efficiency reward against accuracy using a hyperparameter that can be tuned based on your specific needs, whether you prioritise speed or absolute precision.
Results speak louder than words
The experimental results are nothing short of impressive. Testing on DeepSeek-R1-Distill models ranging from 1.5B to 7B parameters across datasets of varying difficulty, AdaptThink delivered consistent wins across the board.
⤠Accuracy boost: +2.3-2.4% improvement over baseline models
⤠Efficiency gains: 40-53% reduction in response length (fewer tokens = lower costs)
⤠Smart adaptation: Up to 99.6% NoThinking usage on appropriate problems
⤠Generalization: +6.5% accuracy on out-of-distribution MMLU benchmark with 38.8% shorter responses
The behaviour matches human intuition beautifully. On easy datasets like GSM8K (basic math word problems), the model predominantly uses NoThinking mode. As problems get harderāmoving to MATH500 and the challenging AIME2024 competition problemsāthe model intelligently shifts toward full reasoning mode. It's like watching an AI develop the same judgment you use when deciding whether to double-check your mental math.

For easier level math problems, AdaptThink clear;y relies on NoThinking mode, and as problems get more challenging, it starts Thinking more often
(source: AdaptThink paper)
What makes these results particularly compelling is that AdaptThink isn't just memorising patternsāit's genuinely learning adaptive reasoning that transfers to completely new domains and problem types.
AdaptThink represents a fundamental shift toward more intelligent AI systemsānot just in raw capability, but in efficiency and practicality. This research directly addresses one of the biggest barriers to deploying reasoning models at scale: cost.
Hereās what we can think of as some immediate real-world use cases for AdaptThink:
Faster user experiences: Chatbots respond instantly to simple queries while maintaining analytical depth for complex requests
Cost reduction: Companies could cut inference costs by up to 50% for typical mixed workloads
Better resource allocation: Educational AI tutors provide quick feedback on basics, detailed reasoning for challenging concepts
Scalable deployment: Advanced reasoning becomes accessible to smaller organisations and applications
AdaptThink feels like one of those research breakthroughs that seem obvious in hindsight but required genuine innovation to achieve. The ability to teach AI when not to overthink might be just as important as teaching it to think deeply in the first place.
How do you see this impacting the AI tools you use daily?
Could adaptive reasoning be the key to making advanced AI both more capable and more accessible?
Share your thoughts with Spark & Trouble.
Want to dive deeper into AdaptThink?
⤠Check out the full research paper
⤠Explore the code & models on GitHub

10x Your Workflow with AI š
Work smarter, not harder! In this section, youāll find prompt templates š & bleeding-edge AI tools āļø to free up your time.
Fresh Prompt Alert!šØ
Ever watched someone blow up on social media seemingly overnight and thought, "How do they DO that?!" while your posts get crickets?
Well, plot twistāthere's a method to the viral madness! This week's Fresh Prompt Alert is your secret weapon to crack the algorithm code. Whether you're building your personal brand or boosting your company's reach, this prompt will help you create content that the algorithms absolutely love.
Time to go viral, friends!
Act as a viral growth hacker. Create a content engine for [niche topic] for [niche audience] that consistently triggers the [platform - LinkedIn/Instagram/TikTok] algorithm and maximizes reach without paid ads.
3 AI Tools You JUST Can't Miss š¤©
Jules: Googleās latest asynchronous Coding Agent
HireHunchās S(ai)na: AI agent that screens the best-fit candidates and can handle all your applications
Routerra: Plan smarter routes in seconds to get more done every day

Spark 'n' Trouble Shenanigans š
Get ready to question everything you see on the internet. Because Google just dropped Veo 3, and honestly, we're not sure if we should be amazed or terrified.
Picture this: You're scrolling through X, and you see a video of a SWAT team infiltrating a terrorist hideout with Hollywood-level dialogue. Your brain goes "wow, cool movie clip!" But plot twistāit's 100% AI-generated. No cameras, no actors, no film crew. Just a prompt and some serious computational wizardry.
Created with Google Flow.
Visuals, Sound Design, and Voice were prompted using Veo 3 text-to-video.
Welcome to a new era of filmmaking.
ā Dave Clark (@Diesol)
9:00 AM ⢠May 21, 2025
Trouble is absolutely buzzing about the technical breakthrough hereāwe're talking hyper-realistic video AND audio generation that's making people do double-takes. Meanwhile, Spark is already scheming about the product implications: at $249.99/month for Gemini Ultra, Google's basically saying "this is premium-tier reality manipulation, folks."
The early adopters are having a field day, flooding social media with everything from Indian chai brewing sessions to college professors explaining Gen Z slang to bewildered boomers. Each clip is so lifelike it's genuinely unsettling. We're watching the birth of a new era where "pics or it didn't happen" just got a whole lot more complicated.
Donāt believe us, check them out for yourselfā¦
šØ This video is made by Veo3, an AI video generator announced by Google.
@GeminiApp#GoogleIO šš¤Æ
ā Indian Tech & Infra (@IndianTechGuide)
2:31 PM ⢠May 21, 2025
A college professor doing a class on Gen Z slang and the video pans over to all the boomers taking notes and seeming super interested #veo3
ā justin (@HonestBlogging)
4:24 AM ⢠May 21, 2025
Before you ask: yes, everything is AI here. The video and sound both coming from a single text prompt using #Veo3 by @GoogleDeepMind .Whoever is cooking the model, let him cook! Congrats @Totemko and the team for the Google I/O live stream and the new Veo site!
ā LĆ”szló GaĆ”l (@laszlogaal_)
7:40 AM ⢠May 21, 2025
Fair warning: Your trust issues with online content are about to get much, much worse.

Well, thatās a wrap! Until then, | ![]() |

Reply