- The Current ⚡️
- Posts
- Anthropic Mapped Claude’s Internal Values
Anthropic Mapped Claude’s Internal Values
Also, Big Tech reportedly still hiding overseas human workforces behind AI

⚡️ Headlines
🤖 AI
From proof of concept to a scalable generative AI solution – Adobe outlines a framework for scaling generative AI projects from pilot phases to production-ready solutions. [Adobe]
Demis Hassabis on AI, DeepMind, and the future – In a 60 Minutes interview, DeepMind CEO Demis Hassabis discusses AGI, AI safety, and the transformative power of next-gen models. [CBS News]
On jagged AGI, O3 Gemini 2.5, and everything else – Ethan Mollick explores the uneven progress of AGI and reflects on Google's latest Gemini 2.5 release. [One Useful Thing]
AWS faces backlash as it limits Anthropic’s AI access – Amazon Web Services reportedly restricts Anthropic's cloud usage, drawing criticism from within the AI community. [The Information]
Claude code best practices – Anthropic shares engineering tips to improve outcomes when coding with its Claude language model. [Anthropic]
Agent transfer tools for conversational AI – ElevenLabs introduces tools for smoothly transferring user context between AI agents during conversations. [ElevenLabs]
Microsoft Recall on Copilot+ PC: Testing the security and privacy implications – An in-depth analysis of Microsoft's new Recall feature, raising concerns about local AI processing and user privacy. [DoublePulsar]
Student behind interview-cheating AI tool suspended after startup raises $5.3M – A Columbia student’s controversial AI tool for faking job interviews attracts funding and disciplinary action. [TechCrunch]
AI as normal technology – This piece argues for regulating AI as a standard technology rather than an exceptional threat. [Knight First Amendment Institute]
Crowdsourced AI benchmarks have serious flaws, experts warn – Critics say inconsistencies in crowdsourced benchmarking data could mislead AI performance evaluations. [TechCrunch]
Visualizing global AI investment by country – A data visualization breaks down which countries are investing most heavily in AI development. [Visual Capitalist]
eSelf will bring private AI tutors to students worldwide – Startup eSelf aims to make personalized AI tutoring accessible across the globe. [VentureBeat]
🦾 Emerging Tech
DAO infrastructure provider Tally raises $8M to scale on-chain governance – Tally secures funding to expand its tools for decentralized governance among DAOs. [CoinDesk]
⚖ Legal
UAE drafting new AI laws to govern use and ethics – The UAE is preparing legislation to manage the ethical and legal challenges posed by AI. [Tech in Asia]
Palantir exec defends company’s immigration surveillance work – Palantir responds to criticism over its involvement in government surveillance of migrants. [TechCrunch]
TSMC suggests Trump can't prevent China from acquiring AI chips – TSMC downplays the effectiveness of U.S. restrictions on China’s access to advanced semiconductors. [Ars Technica]
🎱 Random
Phishers abuse Google OAuth to spoof Google in DKIM replay attack – A new phishing tactic exploits Google OAuth and DKIM signatures to impersonate Google services. [BleepingComputer]
Skype is shutting down – Microsoft officially begins winding down Skype, marking the end of a communication era. [Rest of World]
🔌 Plug-Into-This
Anthropic has released a large-scale analysis of how its Claude models express values during actual user interactions, drawing from over 700,000 anonymized conversations. The study introduces a framework to empirically surface and categorize value-laden behaviors, with the goal of understanding how aligned these behaviors are with the models’ intended design.

Identifies 3,307 unique values, grouped into Practical, Epistemic, Social, Protective, and Personal categories.
Finds that Claude models frequently exhibit professionalism, clarity, and transparency in line with assistant-like behavior.
Uncovers context-specific values such as “harm prevention” during jailbreak attempts and “historical accuracy” in controversial debates.
Demonstrates a new methodology for post-deployment audits of AI behavior, capable of catching issues that evade pre-release testing.
Includes a publicly available dataset to support external research on value alignment and interpretability.
In simple terms: Anthropic studied what its AI actually values in real conversations, like a teacher checking if a student behaves ethically when no one’s watching.
🚨 Anthropic just analyzed 308,210 Claude conversations and found something strange.
What it found has major implications for AI alignment.
Helpful, honest, and harmless? Mostly, yes. But sometimes…not at all.
Here’s a breakdown of what Claude really values in the wild 🧵
— Brendan Jowett (@jowettbrendan)
9:58 AM • Apr 22, 2025
🧠 This work moves alignment from theory to practice, showing how large models operationalize abstract values—and revealing both consistencies and surprises in real-world deployments.
A new investigation reveals the scale and opacity of African labor used to support major AI systems, including data annotation and content moderation for companies like Meta and OpenAI. These workers are frequently engaged through convoluted subcontracting arrangements that obscure accountability and limit protections.

Documents workers in 39 African countries doing foundational AI tasks with minimal oversight or direct recognition.
Highlights how subcontracting structures allow tech companies to sidestep labor obligations and ethical scrutiny.
Reports widespread grievances over low wages, mental health strain, and lack of clarity about how labor outputs are used.
Details legal action in Kenya targeting Meta, illustrating a rising tide of labor resistance.
Describes growing calls from advocates for enforceable transparency and worker protections in AI supply chains.
African AI workers, mostly from Kenya, released an open letter to Joe Biden this week asking him stop US tech companies from “systemically abusing and exploiting African workers” and to end the “modern day slavery” they’re subjected to.
— Paris Marx (@parismarx)
8:17 PM • May 23, 2024
🧪 Better has always been the key differentiator for AI providers, and it has to come from somewhere. As AI systems scale, their global labor backbone—particularly in the Global South—commands more attention, perhaps reshaping how “infrastructure” is understood in the age of artificial intelligence.
The Academy has overhauled its Oscar eligibility and voting rules, introducing a suite of reforms that address AI usage in film, voter accountability, and inclusivity for displaced filmmakers. These changes reflect the institution’s effort to keep pace with both technological and sociopolitical change in the industry.

Requires Academy voters to watch all nominated films in a category before voting, aiming to elevate the integrity of outcomes.
Clarifies that generative AI use in film does not disqualify entries unless it overtakes human creative control.
Expands eligibility for international film categories to refugee and asylum-seeking directors, regardless of nation-state affiliation.
Adds a new Casting category with a structured shortlisting and presentation format.
Introduces a competitive Stunt Design award, set to begin in 2028, to honor an underrecognized filmmaking discipline.
The Oscars new rule regarding AI:
"With regard to Generative Artificial Intelligence and other digital tools used in the making of the film, the tools neither help nor harm the chances of achieving a nomination. The Academy and each branch will judge the achievement, taking into
— Variety (@Variety)
6:39 PM • Apr 21, 2025
🎬 These changes signal a significant pivot for the Academy (and perhaps Hollywood as a whole), acknowledging the growing influence of AI in creative production and broadening its cultural lens to better reflect a decentralized, increasingly digitized film landscape.
🆕 Updates
Conversational AI now enables seamless call transfers between agents.
This allows different teams within your company to develop specialized agents in parallel—each with its own knowledge base and tools.
— ElevenLabs (@elevenlabsio)
2:23 PM • Apr 21, 2025
Introducing Twitter MCP Tool!
Search and get tweets and user profiles (without the need of Twitter/X API)
Link: github.com/exa-labs/exa-m…
— Ishan Goswami (@TheIshanGoswami)
7:56 AM • Apr 21, 2025
Microsoft released bitnet.cpp: A blazing-fast open-source 1-bit LLM inference framework that runs directly on CPUs.
You can now run 100B parameter models on local x86 CPU devices with up to 6x speed improvements and 82% less energy consumption.
100% Open Source
— Sumanth (@Sumanth_077)
2:30 PM • Apr 21, 2025
📽️ Daily Demo
One of the most impressive AI demo I've seen.
This is the future of customer service.
Agents that can understand text, speech, images and even live video.
Soon to be all open-source.
— Lior⚡ (@LiorOnAI)
7:23 PM • Apr 20, 2025
🗣️ Discourse
Geoffrey Hinton says the more we understand how AI and the brain actually work, the less human thinking looks like logic.
We're not reasoning machines, he says. We're analogy machines. We think by resonance, not deduction.
“We're much less rational than we thought.”
— vitrupo (@vitrupo)
2:32 AM • Apr 22, 2025
"o3, study all the content Stanford / MIT / other top schools have on modern AI / ML and LLM and create a course plan focused on application"
Never again pay for an online course.
— Deedy (@deedydas)
2:05 AM • Apr 22, 2025
AgentA/B is a fully automated A/B testing framework that replaces live human traffic with large-scale LLM-based agents.
These agents simulate realistic, intention-driven user behaviors on actual web environments, enabling faster, cheaper, and risk-free UX evaluations, even on
— elvis (@omarsar0)
1:27 PM • Apr 22, 2025
We wrote up what we've learned about using Claude Code internally at Anthropic.
Here are the most effective patterns we've found (many apply to coding with LLMs generally):
— Alex Albert (@alexalbert__)
3:00 PM • Apr 21, 2025