- The Current ⚡️
- Posts
- Anthropic Reveals How Its LLM Claude Thinks
Anthropic Reveals How Its LLM Claude Thinks
Also, Qwen releases QVQ-Max — visual reasoning model

⚡️ Headlines
🤖 AI
Chinese Startup Behind Manus AI Agent Seeks $500 Million Valuation - Manus AI, a Chinese startup claiming to have developed the first general AI agent, is seeking a $500 million valuation amid growing investor interest in AI technologies. [The Information].
ChatGPT is turning everything into Studio Ghibli art — and it got weird fast - OpenAI's "Images for ChatGPT" feature enables users to generate Studio Ghibli-style images, leading to both creative and controversial applications. [The Verge].
Lockheed Martin and Google Cloud Announce Collaboration to Advance Generative AI for National Security - Lockheed Martin and Google Cloud are partnering to develop generative AI solutions aimed at enhancing national security operations. [PR Newswire].
Satya Nadella: DeepSeek is the new bar for Microsoft’s AI success - Microsoft CEO Satya Nadella highlights DeepSeek's R1 model as a new benchmark for the company's AI initiatives, emphasizing efficiency and innovation. [The Verge].
PwC launches AI Agent Operating System for enterprises - PwC introduces 'agent OS,' a platform designed to enable AI agents to communicate and work cohesively within enterprise environments. [PwC].
🦾 Emerging Tech
Unlock the Power of Physical AI with Lenses - Archetype AI introduces 'Lenses,' AI applications built on their Newton™ model, designed to continuously convert raw data into tailored insights for various use cases. [Archetype AI].
GameStop Prices Bitcoin Notes at $29.85 - GameStop has priced its $1.3 billion convertible senior notes, planning to invest the proceeds in Bitcoin as part of its treasury strategy. [CoinDesk].
🤳 Social Media
Expanded conversation ad inventory available to all advertisers with new suitability reporting - Reddit announces expanded conversation ad placements and enhanced brand safety reporting, allowing advertisers to reach users deeper within discussions. [Reddit for Business].
Instagram adds reposts to expand reshare reach - Instagram introduces a repost feature, enabling users to reshare content more easily and expanding the reach of shared posts. [Social Media Today].
🎱 Random
Gaming chat platform Discord in early talks with banks about public listing - Discord is reportedly in early discussions with banks, including JP Morgan Chase and Goldman Sachs, regarding a potential IPO in 2025. [Ars Technica].
Google Maps can soon scan your screenshots to plan your vacation - Google Maps is introducing a feature that identifies locations in users' screenshots, aiding in trip planning by saving and mapping these places. [The Verge].
🔌 Plug-Into-This
Anthropic has introduced a novel methodology for interpreting the internal reasoning pathways of language models, enabling direct visualization of how models process and decompose complex tasks into intermediate steps. Using sparse autoencoders trained on internal activations, the team has begun identifying discrete, interpretable features that resemble "thought fragments."

The technique applies unsupervised sparse autoencoders to the activations of Claude 2.0 models, yielding millions of latent features.
Many features correspond to semantically rich concepts, such as tone (e.g. "politeness") or task structure (e.g. "step-by-step reasoning").
These features can be "ablated" or removed to test causal influence on downstream model behavior, validating their functional relevance.
The work suggests that certain model capabilities—like chain-of-thought reasoning—emerge as compositional structures across latent dimensions.
Early results raise the possibility of intervening in specific cognitive patterns inside LLMs, potentially opening a path to more steerable and auditable AI systems.
New Anthropic research: Tracing the thoughts of a large language model.
We built a "microscope" to inspect what happens inside AI models and use it to understand Claude’s (often complex and surprising) internal mechanisms.
— Anthropic (@AnthropicAI)
5:00 PM • Mar 27, 2025
🧠 By surfacing latent units tied to abstract reasoning modes, this approach parallels neuroscience-style feature mapping, edging LLM interpretability closer to a functional cartography of thought—without requiring ground-truth labels or intrusive retraining.
Alibaba’s Qwen team has unveiled Qwen-QVQ-Max, an advanced vision-language model architecture featuring a 32K context window and a refined image encoder-decoder stack. Optimized for spatially dense tasks such as document understanding, chart analysis, and grounded reasoning, QVQ-Max introduces major upgrades to both visual input handling and long-context alignment.

The model architecture introduces a ViT-based image encoder paired with a quantized vision-query decoder to enhance multi-scale visual representation.
QVQ-Max supports input images up to 896×896 resolution and can process multi-image prompts within a 32K token window.
It integrates a “QVQ” mechanism—quantization-vision-query—that improves visual grounding and region-level reasoning over complex layouts.
Benchmarks show marked gains over previous Qwen-VL models in ChartQA, DocVQA, and multi-hop visual question answering.
The system demonstrates robust performance in real-world applications like invoice understanding, slide parsing, and web UI interpretation.
✨ Excited to share QVQ-Max, our visual reasoning model that's still evolving
We've been experimenting with this approach for a while - try it out on Qwen Chat! (chat.qwen.ai) 🚀
Just upload any image or video, ask away, and hit the "Thinking" button to see how it— Qwen (@Alibaba_Qwen)
7:32 PM • Mar 27, 2025
👁️ What sets QVQ-Max apart is its architectural emphasis on visual quantization and query composition—moving beyond generic image encoders to a system where visual regions are discretized into learnable tokens, enabling precise, referenceable grounding across long contexts. This elevates multimodal reasoning from perceptual matching to structured visual abstraction.
Databricks is pioneering a label-free approach to large language model fine-tuning by leveraging proprietary retrieval systems and advanced synthetic data pipelines. Their new platform, aimed at enterprise users, integrates RAG-enhanced data sourcing with open LLMs to boost accuracy, reduce hallucination, and preserve data governance.

The method centers on "instruction-following" tuning using domain-specific documents without the need for human annotations.
Databricks employs RAG to selectively pull high-relevance documents, which are then used to generate synthetic instruction datasets.
Fine-tuning is performed on open models such as Mistral or LLaMA variants within a governed enterprise environment.
This workflow supports vertical-specific applications—e.g. legal, finance, biotech—where labeled data is scarce or sensitive.
The system is deployed within Databricks’ AI/BI stack, facilitating unified access to data, model customization, and visualization tools.
Explore the latest releases and enhancements in Databricks SQL! Including:
- Accelerated data migrations with BladeBridge
- AI/BI Dashboard and Genie updates
- Git support for queriesGet the full rundown👇
dbricks.co/4iEHjOf— Databricks (@databricks)
5:00 PM • Mar 27, 2025
🔧 The approach sidesteps costly manual labeling by orchestrating in-house retrieval and synthesis, underscoring Databricks’ strategy of anchoring generative AI in enterprise-native data flows rather than chasing model scale.
🆕 Updates
it's super fun seeing people love images in chatgpt.
but our GPUs are melting.
we are going to temporarily introduce some rate limits while we work on making it more efficient. hopefully won't be long!
chatgpt free tier will get 3 generations per day soon.
— Sam Altman (@sama)
4:32 PM • Mar 27, 2025
BREAKING 🚨: Anthropic is preparing to release a new version of Claude Sonnet 3.7 with a 500K Context Window (Current is 200K).
— TestingCatalog News 🗞 (@testingcatalog)
11:24 PM • Mar 26, 2025
BREAKING 🚨: ChatGPT can now connect to your org’s Google Drive workspace, pulling in internal knowledge in real-time to provide more personalized responses.
It will be rolling out gradually to Team customers over the next few weeks.
— TestingCatalog News 🗞 (@testingcatalog)
7:57 PM • Mar 27, 2025
📽️ Daily Demo
In the past 1 month, I've made 9 game prototypes with Grok, Cursor and Windsurf.
Here's how I usually start my initial prompts:👇
— Danny Limanseta (@DannyLimanseta)
2:23 PM • Mar 26, 2025
🗣️ Discourse
In 2016, Studio Ghibli staff pitched co-founder Hayao Miyazaki on a rough AI model that could be used for future animation projects
His reaction: "I strongly feel that this is an insult to life itself."
— Morning Brew ☕️ (@MorningBrew)
2:01 PM • Mar 27, 2025
ChatGPT 4o's new image gen is insane. Here's what Severance would look like in 8 famous animation styles 🧵
1/8:
Rankin/Bass – That nostalgic stop-motion look like Rudolph the Red-Nosed Reindeer. Cozy and janky.— Bennett Waisbren (@BennettWaisbren)
1:17 PM • Mar 27, 2025
Created the image with ChatGPT. Brought it to life with Hedra.
Just upload a photo, drop in your script, and hit generate.
That’s it. Video creation has never been this easy—or this accessible.
— Hedra (@hedra_labs)
3:18 PM • Mar 27, 2025