
Pressure is building in places most teams do not instrument. AI budgets are climbing, deployment cycles are tightening, and toolchains are expanding faster than anyone can govern them. On the surface, it reads like acceleration. Underneath, it is fragmentation with better branding. Systems that were already misaligned are now being stress tested in real time, and the results are starting to show.
That tension is exactly where Jellyfish steps in with “Platform Engineering in the AI-Native Era: How to Scale AI Adoption Without Chaos,” landing Mar 19, 2026 at 2:00 PM ET as a 1 hour webinar that feels less like a session and more like a systems check. The premise is blunt and earned. AI doesn’t fix broken systems. It magnifies them. And right now, most organizations are somewhere between experimentation and something they are not yet ready to call integration.
Inside this room, virtual but far from passive, the conversation tightens around what happens when AI stops being a side project and starts becoming infrastructure. Not theory. Not vibes. Real operators dealing with real scale, where every new model, tool, or workflow either compounds leverage or compounds confusion.
Ryan Erickson brings the practitioner lens from SPS Commerce, where developer productivity is a measurable constraint tied to output and clarity. Sam Barlien carries the ecosystem view from Platform Engineering, connecting the dots across companies trying to turn internal platforms into something developers actually use instead of work around. Liz Coolman and Alexa Lytle step in from Jellyfish with product depth, translating engineering activity into something the business can understand, track, and act on without guesswork.
The throughline is platform engineering, but not as a buzzword people throw into decks to sound current. This is about golden paths that people follow because they work, not because they are mandated. Automation that reduces decisions instead of multiplying them. Developer portals that feel less like documentation and more like direction. And the uncomfortable but necessary question of measurement, because if 86% of organizations say platform engineering is essential to realizing AI’s business value, someone has to prove where that value actually shows up.
Jellyfish is not just hosting the conversation. Jellyfish is framing the scoreboard. Platform adoption, AI impact, developer experience, all pulled into view so leaders can see what is moving and what is just making noise. The market does not need another AI success story. It needs fewer invisible failures, and a clearer line between building with AI and actually operating it.