<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[Bedside Brief]]></title><description><![CDATA[Monthly notes on AI in advanced care at home]]></description><link>https://brief.bedsidehealth.com</link><image><url>https://substackcdn.com/image/fetch/$s_!9wM9!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe0be75fa-73cd-456a-a0b0-892e4beb3783_2048x2048.png</url><title>Bedside Brief</title><link>https://brief.bedsidehealth.com</link></image><generator>Substack</generator><lastBuildDate>Fri, 15 May 2026 09:45:42 GMT</lastBuildDate><atom:link href="https://brief.bedsidehealth.com/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Oliver Chan]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[bedsidebrief@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[bedsidebrief@substack.com]]></itunes:email><itunes:name><![CDATA[Bedside Brief]]></itunes:name></itunes:owner><itunes:author><![CDATA[Bedside Brief]]></itunes:author><googleplay:owner><![CDATA[bedsidebrief@substack.com]]></googleplay:owner><googleplay:email><![CDATA[bedsidebrief@substack.com]]></googleplay:email><googleplay:author><![CDATA[Bedside Brief]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[Notes from this morning's panel on agentic AI with Stanford]]></title><description><![CDATA[Three insights on agentic deployments at Stanford, and two more from across the field.]]></description><link>https://brief.bedsidehealth.com/p/notes-from-this-mornings-panel-on</link><guid isPermaLink="false">https://brief.bedsidehealth.com/p/notes-from-this-mornings-panel-on</guid><dc:creator><![CDATA[Bedside Brief]]></dc:creator><pubDate>Wed, 13 May 2026 22:53:36 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!9wM9!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe0be75fa-73cd-456a-a0b0-892e4beb3783_2048x2048.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>One more thanks to <a href="https://www.linkedin.com/in/kameron-black/">Kameron Black</a> and <a href="https://www.linkedin.com/in/timothy-keyes/">Timothy Keyes</a> for a panel that named limits as well as the wins. Key highlights:</p><p><strong>1. Eligibility screening is no longer a research problem, it&#8217;s a buildable one. </strong>Stanford&#8217;s surgical co-management agent has been in production for six months. <a href="https://arxiv.org/abs/2603.17234">SCM Navigator</a> triages incoming surgical patients for hospitalist consult. Across 6,193 patients, it runs at 83.6% accuracy in deployment, drives ~90% of the SCM team&#8217;s caseload, and has been associated with zero patient safety events. For HaH programs filtering hundreds of patients for candidacy each day against well-established criteria, this is the same problem, and the solution isn&#8217;t theoretical anymore.</p><p><strong>2. Conversely, agents work &#8212; just not for every job.</strong> On long-horizon tasks in <a href="https://arxiv.org/abs/2605.02240">PhysicianBench</a>, GPT-5.5 performed best, but got it right on the first try only 46% of the time. The bulk of failures came from clinical reasoning itself, not from data lookup or EHR actions. But on bounded tasks, they ship. Timothy summed up Stanford's deployment record: <em>"We've had a lot of success with deployments, even those where the level of autonomy is actually relatively low."</em> The opening for HaH leaders building today: scope deployments to what agents are actually good at &#8212; screening, drafting, retrieval, surfacing decompensation signals &#8212; and keep the clinical judgment with your team. That's where your program's edge is.</p><p><strong>3. The model is the easy part. Your local context is the moat.</strong> Stanford&#8217;s <a href="https://arxiv.org/abs/2501.14654">MedAgentBench V2</a> showed that better tool design, hospital-specific knowledge fed in as worked examples, and access to a memory of prior cases raised performance from 70% to 91% on the same benchmark. The win there isn't model selection; it's tool design, hospital-specific worked examples, and accumulated case memory. Kameron framed where this leads: <em>"Memory will enable specialty preferences and institutional protocols to be baked in."</em> If you're building, this is the part your team actually owns: how your program runs, the cases you've seen, the workflow you've refined. The frontier model is a commodity, but your institutional know-how is not.</p><h2>Two more from across the field</h2><p><strong>Anthropic: assemble agent infrastructure like an operating system, not one all-in-one app.</strong> Their April <a href="https://www.anthropic.com/engineering/managed-agents">engineering essay</a> walks through what they got from splitting the parts that *think*, *act*, and *remember* into independent, replaceable pieces &#8212; each free to evolve as the underlying models improve. They saw a 90% drop in their slowest response times after the restructure.</p><p>For HaH leaders thinking about how to build: modular wins. The constant pace of model improvement from frontier labs is the largest free source of gains, so design your agent system to capitalize on it, not compete.</p><p><strong>MGB is betting expansion on eligibility automation. </strong>Mass General Brigham&#8217;s HaH program <a href="https://www.beckershospitalreview.com/healthcare-information-technology/innovation/mass-general-brigham-enters-growth-mode-after-cms-hospital-at-home-waiver-extension/">announced</a> entering &#8220;growth mode&#8221; &#8212; now five hospitals, 70-bed capacity, expanding into oncology, postoperative care, behavioral health, and dementia. Capacity growth was explicitly tied to ML-driven eligibility screening and a Philips partnership on an &#8220;<a href="https://www.massgeneralbrigham.org/en/about/newsroom/articles/himss-2026-shaping-the-future-of-healthcare">autonomous digital workforce</a>&#8221;. </p><p>The largest HaH program in the country is treating eligibility automation as a build, not a side experiment, and tying it directly to how they grow.</p>]]></content:encoded></item></channel></rss>