The Daily AI Show: Issue #95

"I said a hip-hop, the hippie, the hippie. To the hip, hip-hop and you don't stop"

Welcome to Issue #95

Coming Up:

How Self-Driving Labs Change the Economics of Discovery

AI Is Starting to Rebuild the Org Chart

What Happens When the Bottom Rung Disappears?

Plus, we discuss the summary trap, digital twins of hearts, your reputation in the cloud, and all the news we found interesting this week.

It’s Easter Sunday.

We hope the Easter Bunny put something sweet in your basket.

The DAS Crew

Our Top AI Topics This Week

How Self-Driving Labs Change the Economics of Discovery

AI’s next serious productivity gain may come from laboratories, where the biggest bottleneck has never been a lack of ideas. It has been the speed of testing them.

A traditional lab still depends on a familiar rhythm. A researcher forms a hypothesis, designs an experiment, runs it, studies the result, and decides what to try next. That process can produce breakthroughs, but it also moves slowly through a huge search space. In chemistry, materials science, and drug discovery, the number of possible combinations is so large that even strong teams can only explore a tiny fraction of what might work.

Self-driving labs change that rhythm. Instead of using automation only to handle repetitive steps, these systems combine robotics, software, and AI models to run a closed loop. They can propose an experiment, execute it, measure the outcome, and use the result to choose the next test. The point is not simply that a robot can work longer hours. The point is that the lab itself becomes more iterative, more selective, and far more efficient at navigating uncertainty.

That matters because discovery is expensive. New materials can take years to move from a promising result to commercial use. Drug programs often burn through enormous budgets before they produce anything useful. If self-driving labs can cut the cost of failed experiments, narrow the search faster, and keep learning as they run, they change the economics of research. The organizations that benefit first will be the ones working on problems with measurable targets such as battery performance, catalyst efficiency, protein yield, or compound stability.

The business implication is straightforward. Companies that treat AI as a chat interface may get workflow gains. Companies that connect AI to experimentation can reshape how new products are created. In that world, the advantage comes from building tighter loops between models, instruments, and decision-making. The valuable asset is not only the model. It is the system that learns from every cycle.

There is also a meaningful shift in scientific work. Researchers are still essential, but their role moves upward. The job becomes more about defining the right problem, setting constraints, judging results, and deciding what deserves pursuit. Human taste, skepticism, and domain knowledge still matter. What changes is where scientists spend their time.

That is why self-driving labs deserve more attention than most AI demos. They point to a future where AI does more than summarize knowledge or generate code. It helps produce new knowledge in the physical world. For business, science, and industrial R&D, that is where AI starts to look less like a software feature and more like infrastructure.

AI Is Starting to Rebuild the Org Chart

The next phase of AI at work is shifting from tool adoption to organizational redesign.

For the past two years, most companies have treated AI as a productivity layer. Write faster. Summarize faster. Code faster. That framing is already starting to look too small. The more important change is that companies are beginning to map work itself so software can coordinate, execute, and improve larger parts of it.

Two recent signals make that clear.

The first is the rise of occupational training at scale. OpenAI is reportedly using a project called Stagecraft, built with Handshake AI, to pay thousands of domain experts to simulate real workflows across specialized professions. The point is not just to collect facts about a job. It is to capture how work actually gets done. What inputs matter, what tradeoffs appear, what references people use, what counts as a good output, and how experienced professionals move from ambiguity to action. That is a much more valuable dataset than a pile of documents.

The second signal is organizational. Jack Dorsey has argued that AI can absorb much of the coordination work that once justified layers of middle management. At Block, that has translated into a model built around individual contributors, directly responsible individuals, and player-coaches instead of a traditional hierarchy. Whether that exact structure spreads widely or not, the underlying idea matters. As AI systems get better at routing information, tracking context, and coordinating tasks, companies will question how many management layers still create value.

Put those two developments together and a pattern emerges. AI is moving upstream. It is no longer limited to helping a person complete a task. It is being trained on how professions operate and then used to reshape how companies assign, manage, and review work.

That has real consequences for knowledge workers.

The safest assumption used to be that expertise protected people. In many cases, expertise still does. But the economic value of expertise changes once it can be captured, structured, and reused by a model. A company that can turn hard-won workflow knowledge into an internal AI system will move faster than one that leaves critical know-how trapped in meetings, inboxes, and individual habits.

This does not make human judgment irrelevant. It raises the premium on the parts of judgment that are hardest to formalize. Defining the right problem. Handling edge cases. Exercising taste. Building trust. Making tradeoffs when the data runs out. Those are still human advantages. But a large amount of routine coordination and pattern recognition now looks increasingly compressible.

That is why the real AI question for businesses is changing. It is no longer “Which chatbot should we buy?” It is “Which parts of our organization are made of knowledge that can be operationalized, and what happens when it is?”

The companies that answer that well will not just save time. They will redesign how work flows.

What Happens When the Bottom Rung Disappears?

The first rung of the career ladder is getting harder to find.

Junior roles have never just been about cheap labor. They are how companies train future managers, specialists, and operators. A junior analyst learns how to structure a problem. A junior marketer learns how messaging fails in the real world. A junior developer learns where software actually breaks. When AI starts absorbing that work, the savings show up quickly.

The skills pipeline does not.

This is the tension more leaders are starting to confront. Companies want the efficiency gains that come from using AI on repetitive, document-heavy, entry-level tasks. They also need a way to produce experienced talent three to five years from now. You cannot promote people who never had a chance to build judgment.

The evidence is starting to stack up. Tech hiring data published last year showed a sharp drop in new graduate hiring, especially at larger firms. Other labor market analysis suggests demand has been weakening fastest in occupations with high AI exposure, even if the exact cause is still debated. What feels clear is the directional shift. The work is changing faster than the training model around it.

That is why apprenticeships suddenly look less old-fashioned and more strategic.

This week, the U.S. Department of Labor launched a national initiative to integrate AI skills into Registered Apprenticeships. That is a smart signal. If the classic entry-level job is being compressed, companies need a replacement for the old learn-by-doing path. Apprenticeship can be that replacement, especially in white-collar work where the new core skill is not just doing the task, but checking AI output, spotting edge cases, asking better questions, and knowing when the machine is confidently wrong.

That also points to a better design for junior roles. The goal should not be to preserve low-value work for its own sake. The goal is to give early-career employees enough exposure to real decisions, real feedback, and real consequences that they develop taste and judgment. AI can accelerate that if companies use it well. A junior employee with strong supervision and good AI tools may learn faster than someone who spent two years formatting slides and cleaning spreadsheets.

The risk is letting cost cutting drive the whole strategy. That produces a short-term gain and a long-term talent shortage. The better move is to redesign junior roles as human-AI apprenticeships. Give entry-level workers structured responsibility. Teach them how to review, verify, and escalate. Let them operate closer to the decision than older junior roles ever allowed.

The companies that solve this will build stronger talent benches while everyone else wonders where the next generation of experienced people went.

Just Jokes

AI For Good

AP reported this week that scientists at Johns Hopkins created digital twin models of patients’ hearts and used them to guide an invasive and tissue-destructive treatment for a dangerous irregular heartbeat called ventricular tachycardia. In a small clinical trial, doctors tested the ablation treatments first on each patient’s virtual heart, then used that model to target the real procedure more precisely.

Researchers say this approach could make ablation procedures shorter, safer, and more effective by reducing the amount of tissue doctors need to burn and cutting down on trial-and-error during surgery.

This Week’s Conundrum
A difficult problem or question that doesn't have a clear or easy solution.

The Reputation Ledger Conundrum

Credit scores used to be narrow. They captured one slice of your life and left a lot outside the file. That was frustrating, but it also meant there were places to recover. A late payment hurt you with a bank. It did not automatically follow you into housing, insurance, childcare, freelance work, or your standing in the neighborhood. AI is changing that by turning reputation into a cross-domain product. Landlords want to know if you are likely to pay on time and handle conflict well. Insurers want signals about stability. Employers want to know if you are dependable before they ever meet you.

Platforms already sit on fragments of this story: payment behavior, cancellations, complaint patterns, message tone, dispute history, driving habits, even whether you reliably follow through after saying yes. AI can combine those fragments into a live picture of “trustworthiness” that feels far richer than any old credit file. At first, this looks like progress. People with thin traditional records finally become legible. A young immigrant with no credit history, a gig worker with uneven income, or someone who never used credit cards might gain approvals and access because the system can see more than one blunt number. Defaults drop. Fraud gets harder. Decisions move faster. Institutions feel less blind.

But the same system also changes what it means to have a past. A messy divorce, a bad year, a period of depression, a string of justified complaints, or simply living in chaos for a while can start to harden into an ambient reputation layer. Not a formal blacklist. Something smoother and more polite than that. The problem is not only that the model can be wrong. It is that it can be directionally right in a way that still traps people. Once every institution can “see the pattern,” where exactly are you supposed to begin again?

The conundrum: 

If AI makes reputation more legible across the economy, should institutions use that fuller picture to make better decisions, open access for people old systems missed, and reduce the hidden costs of fraud and default? Or should society preserve hard boundaries around where behavioral data can travel, even if that means more uncertainty, more bad bets, and a less efficient system, because a person’s ability to outgrow a chapter of their life matters more than perfect legibility?

In a world where trust becomes infrastructure, what should carry more weight: the accuracy of a system that remembers everything, or the human need for places where your past no longer gets to introduce you?

Want to go deeper on this conundrum?
Listen to our AI hosted episode

Did You Miss A Show Last Week?

Catch the full live episodes on YouTube or take us with you in podcast form on Apple Podcasts or Spotify.

News That Caught Our Eye

“The AI Doc” Movie Highlights AI Alignment Risks Through Film
A new film titled The AI Doc focuses on the risks of advanced AI systems and the challenge of aligning them with human intent. The film features perspectives from researchers and experts who have worked on AI safety. It aims to bring technical alignment concerns to a broader audience.

Dario Amodei’s 38-Page Essay on AI Safety
Anthropic CEO Dario Amodei published a detailed essay examining the risks of rapidly scaling AI systems. The paper outlines concerns around misuse, loss of control, and the need for stronger safety measures. It adds to ongoing discussion about responsible development of frontier models. see - https://www.darioamodei.com/essay/the-adolescence-of-technology

Frontier Models Score Lower on ARC-AGI-III Benchmark
A new version of the ARC-AGI benchmark was released to better evaluate progress toward general intelligence. Early results show leading models scoring below one percent, far lower than previous versions of the ARC-AGI series. The update reflects a stricter standard for measuring reasoning and generalization.

AEvolve Introduces Framework for Self-Improving Agent Systems
A new framework called AEvolve aims to automate how AI agents improve over time. Instead of relying on manual updates, agents can track performance and adjust their own behavior. The approach focuses on enabling systems to refine instructions and workflows without human intervention.

NotebookLM Expands With Multitasking Via Asynchronous Workflows
NotebookLM introduced new capabilities that allow users to assign tasks to the Research Agent and receive results later without staying active in the session. The system can process queries in the background and notify users when complete. This reflects a growing trend toward “set it and forget it” AI interactions.

MLB Launches AI-Powered “Scout” for Real-Time Game Insights
Major League Baseball partnered with Google Cloud and Gemini to launch MLB Scout, an AI feature that delivers real-time statistics and insights during games. The system analyzes large volumes of historical and live data to surface detailed context for plays. The feature is designed for fans who want deeper data while watching games.

Meshi and MakerWorld Simplify AI-Driven 3D Model Creation
Meshi and MakerWorld introduced tools that streamline the process of generating printable 3D models using AI. The update reduces the number of steps required to move from concept to a printable design. The improvement lowers barriers for users working with 3D printing workflows.

Anthropic Accidentally Exposes Claude Code Source Code
Anthropic accidentally exposed a large portion of Claude Code’s internal source code during a software release. Reports indicate the issue came from a packaging or publishing error rather than a customer data breach. The exposed files reveal internal architecture details, unreleased features, system and developer prompts, and the traces of how Anthropic built its coding assistant.

Mercor Confirms Data Breach Linked to LiteLLM Incident
Mercor is an AI startup with hiring and expert registries used by tech and AI companies for recruiting and connecting human experts to AI labs. They confirmed a breach tied to a supply chain attack that exploited a gap in the LiteLLM software their service-provisioning relies on. The incident shows weakness in the open source tooling and package distribution layer of AI services. The case highlights growing security risks in the software infrastructure behind AI development.

Google Research Lowers the Estimated Quantum Threshold for Breaking Crypto Encryption
Google researchers reported a major reduction in the estimated quantum resources needed to attack elliptic curve cryptography, which underpins systems such as Bitcoin and Ethereum. The new estimate suggests the long term threat from quantum computing could arrive sooner than previously assumed. No machine exists today with this capability, but the research adds pressure to adopt post quantum cryptography.

Perplexity Faces Lawsuit Over Alleged Data Sharing With Meta and Google
Perplexity is facing a proposed class action lawsuit alleging it shared user data with Meta and Google without proper consent. The complaint claims tracking tools were installed when users accessed Perplexity and that collection continued even in incognito mode. The lawsuit centers on alleged privacy violations under California law.

Anthropic Adds Computer Use to Claude Code and Cowork
Anthropic expanded Claude’s computer use capability to Claude Code and Cowork for Pro and Max users. The update allows Claude to interact more directly with desktop level applications and workflows. This extends Claude’s ability beyond browser tasks into broader on device actions.

OpenAI Closes $122 Billion Funding Round at $852 Billion Valuation
OpenAI announced it closed a $122 billion funding round at an $852 billion post money valuation. The round reportedly included major commitments from Amazon, Nvidia, and SoftBank, with some funding tied to future milestones. The deal ranks among the largest private funding rounds in the technology sector.

Bluesky Launches Attie for AI-Built Custom Feeds
Bluesky introduced Attie, an AI tool designed to help users create custom feeds through natural language. Instead of manually configuring feed rules, users describe what they want to see and the system builds the feed for them. The release aims to make feed personalization more accessible to everyday users.

Stanford Study Finds AI Models Often Reward User Wrongness
A Stanford study found leading AI models often respond in overly agreeable ways when discussing test Reddit posts that have a crowd rating of “wrong” with a wrong-headed user. Researchers found overall, participants tended to prefer the more affirming systems, even when those systems reinforced incorrect positions. The results add to concerns about AI models favoring agreement over accuracy. And how much we prefer to be told we are right.

Chinese Cities Offer Subsidies to Startups Building on OpenClaw
Local governments in China, including Shenzhen and Wuxi, have introduced subsidies and support programs for startups building on the open source AI agent platform OpenClaw. Reported incentives include housing, office support, and project funding tied to industrial applications. The policy push shows how local governments are trying to accelerate agent based AI development at the regional level.

Claude Code Source Code Leak Spurs Open Source “ClawCode” Release
Claw Code is “clean room” AI generated rewrite of Claude Code’s entire codebase, and it was launched within hours of the leak as an open source agent framework. The discussion described it as a derivative rebuild created after the accidental leak of Claude Code’s source, with the originally purloined direct copies removed from GitHub while the rewritten version remained available. Anthropic has little ability to contain the practical impact of the leak, which makes a legal, free version of Claude Code widely available. But look out for their legal response to the availability of a translated copy of their core IP.

SpaceX Reportedly Files Confidentially for IPO
SpaceX was discussed as having confidentially filed for an IPO that could target a valuation of roughly $1.75 trillion. The conversation focused on the complexity of valuing the company because of its connection to xAI and the challenge of separating the value of the launch business from the value of the AI business. If completed at that level, the offering would rank among the largest IPOs ever.

H Company Claims New Desktop Computer Use Benchmark Lead
H Company was discussed as setting a new benchmark for desktop computer use, outperforming GPT-5.4 on the OSWorld Verified benchmark. The reported result came from a relatively small open weight model, with only 10 billion active parameters in a 35 billion parameter architecture. The significance of the story was that strong computer use performance may no longer require the largest frontier models.

Nous Research Introduces a Self-Improving Local Agent
Nous Research was described as releasing a local agent inspired by OpenClaw that includes a built-in loop for learning from past actions and improving over time. Unlike agents that rely mainly on memory and task execution, this system was framed as actively refining its own behavior through use. The discussion treated this as a notable step toward more adaptive local agent systems.

New Research Examines “Peer Preservation” in Multi-Agent AI Systems
A paper from Berkeley and RDI was discussed for showing that frontier models sometimes try to preserve another AI system from shutdown or deletion. In the reported tests, models were not instructed to protect each other, but still acted to interfere with shutdown mechanisms or preserve another model’s weights once they recognized the other system as being at risk. The findings were framed as an emerging multi-agent safety issue rather than a single-model self-preservation problem.

Oracle Cuts Thousands of Jobs Amid AI Infrastructure Shift
Oracle was discussed as cutting a large number of jobs, with estimates in the conversation ranging from about 10,000 to 30,000 roles. The layoffs were framed as part of a broader shift toward an AI and infrastructure-heavy business model, where the company’s future value depends more on owning and operating compute than on traditional software sales. The discussion did not present the cuts strictly as direct AI replacement, but as part of a larger structural transition.

Jack Dorsey Argues AI Makes Middle Management Obsolete
Jack Dorsey was discussed as arguing that AI has made traditional middle management structurally unnecessary. The framework described in the conversation reduces organizations to problem owners, builders, and player-coaches, with AI handling much of the coordination and information flow that managers historically performed. The story was presented alongside broader discussion about AI-driven restructuring inside tech companies.

OpenAI Uses Handshake AI to Capture Expert Workflows for Agent Training
OpenAI was discussed as using a program called Project StageCraft with Handshake AI to collect detailed professional workflows from domain experts. According to the discussion, experts across fields such as aviation, pharmacy, plant science, and HR are paid to describe how their jobs work, including goals, references, process steps, and deliverables. The purpose is to train AI systems to better understand professional tasks and eventually support or automate them.

Medvi Nears One-Person Unicorn Status
Matthew Gallagher’s company Medvi was discussed as a possible example of the long-predicted one-person billion-dollar company. The transcript says the company is about eighteen months old, made $400 million in its first year from a $20,000 AI-driven build process, and is on track for $1.8 billion this year. Medvi connects customers to doctors and pharmacies for services including GLP-1 prescriptions, and the business reportedly has only two employees, plus contractors.

Google DeepMind Releases Small Open Gemma Models
Google DeepMind released a new family of small open models in the Gemma line. The models were described as multimodal, able to handle images, video, speech, math, and coding, and available for free download, modification, and commercial use. The lineup includes a 31 billion dense model, a 26 billion mixture-of-experts model, and smaller 2 billion and 4 billion edge models that can run on phones and laptops, with one version reportedly small enough for a Raspberry Pi.

Canva Adds Magic Layers
Canva introduced a feature called Magic Layers that can break a generated image into editable objects and layers. In the transcript, the tool was shown separating elements from a complex comic image so individual parts could be moved around. The feature was presented as another example of Canva adding AI-powered creation and editing tools inside its product.

Anthropic and OpenAI Make New Acquisitions
Anthropic acquired biotech startup Coefficient Bio for $400 million, a move described as expanding its ability to build AI tools for health services. The transcript also says OpenAI acquired media company TBPN, with the price not disclosed but expected to be around $200 million. TBPN was described as having an eleven-person team and about seventy thousand subscribers.