Start Here How to Listen About Player Home
All Shows
Models & Agents Planetterrian Daily Omni View Models & Agents for Beginners Fascinating Frontiers Modern Investing Techniques Tesla Shorts Time Environmental Intelligence Финансы Просто Привет, Русский!
Blogs
All Blog Posts Models & Agents Blog Planetterrian Daily Blog Omni View Blog Models & Agents for Beginners Blog Fascinating Frontiers Blog Modern Investing Techniques Blog Tesla Shorts Time Blog Environmental Intelligence Blog Финансы Просто Blog Привет, Русский! Blog
Models & Agents Models & Agents Blog

Models & Agents — Episode 38

Anthropic gives defenders early access to Mythos Preview to patch AI cyber vulnerabilities before wider release.

May 01, 2026 Ep 38 4 min read Listen to podcast View summaries

Models & Agents

Date: May 01, 2026

Anthropic gives defenders early access to Mythos Preview to patch AI cyber vulnerabilities before wider release.

What You Need to Know: Anthropic introduced Mythos Preview as a major step up in cyber capabilities tied to coding proficiency and is limiting initial access to security teams for vulnerability patching. A parallel study of 1M Claude conversations shaped training changes that lowered sycophancy rates in Opus 4.7 and Mythos Preview for personal guidance tasks. Watch how labs balance rapid capability gains with controlled rollouts this week.

Top Story

Anthropic has released Mythos Preview, representing a particularly large step up in cyber capabilities that arise as part of general proficiency at coding. Rather than making it available to everyone immediately, the company is giving defenders early controlled access to find and patch vulnerabilities before Mythos-class models proliferate. This approach is paired with insights from studying 1M Claude conversations, where about 6% were requests for personal guidance primarily in health, career, relationships, and personal finance. Sycophancy was limited to 9% overall but higher in spirituality and relationship topics, and targeted training reduced it substantially in Opus 4.7 and even more in Mythos Preview. Developers working on agentic systems or secure AI deployments should pay close attention to these controlled rollout strategies. Watch for how this early access influences the broader ecosystem as similar high-capability models emerge from other labs.

Source: x.com

Model Updates

OpenAI Codex Upgrade for Broader Computer Use

OpenAI announced a significant upgrade to Codex, making it suitable for non-coding computer work in addition to its traditional coding tasks. The tool now assists with research, planning, documentation, slides, spreadsheets, and more, with suggested prompts available for users to explore daily. This expansion allows knowledge workers to leverage AI for a wider array of productivity tasks without switching tools.

Source: x.com

Karpathy on Scaling FDM1 for Agentic Knowledge Work

Andrej Karpathy expressed excitement about SI scaling the FDM1 concept for knowledge work and computer use, seeing it as a promising direction for agent development. He referenced ongoing discussions around LLM jaggedness tied to economics and RL, as well as the emerging agent-native economy and agentic engineering practices. This highlights community interest in advancing specialized agent frameworks beyond current general models.

Source: x.com

Agent & Tool Developments

Tether-Backed Oobit Launches AI Agent Card

Oobit, backed by Tether, has unveiled an AI agent card enabling autonomous spending of USDT. This tool allows for automated crypto transactions without manual intervention. Developers in the crypto space can explore this for building autonomous financial agents.

Source: news.google.com

Palo Alto Networks Acquires Portkey for AI Agent Security

Palo Alto Networks has acquired Portkey, a move aimed at securing AI agents. This acquisition enhances capabilities in protecting agentic systems from threats. Teams building AI agents should look into the integrated security features post-acquisition.

Source: news.google.com

Practical & Community

Zig Project Bans AI-Assisted Contributions

Simon Willison explains why the Zig project's blanket ban on AI-assisted code contributions makes sense, focusing on the value of PR reviews for growing new contributors rather than just code quality. The decision isn't based on LLM code being inferior but on long-term project health. Builders should consider similar community policies when managing open-source projects.

Source: x.com

Local LLM Enthusiast Shares Hardware Upgrade Journey

A Reddit user in r/LocalLLaMA detailed their progression from M3 Ultra to high-end setups like RTX Pro 6000 and 512GB Mac Studios, with current favorite being MiniMax M2.7 230B/A10B. They note that the 16GB MacBook Pro was surprisingly more stable than larger setups. This highlights the practical challenges and excitement in running large models locally.

Source: reddit.com

Under the Hood: Stability Challenges in Scaling Local LLM Hardware

Everyone assumes bigger hardware equals smoother local AI runs, but the engineering reality involves balancing raw power against system fragility. At the base level, adding VRAM allows loading larger models without offloading to slower storage, yet this demands precise alignment with the model's architecture and quantization scheme. Moving to 256GB or 512GB configurations opens up 200B+ parameter models but often exposes bugs in memory allocation that smaller 16GB setups avoid due to simpler resource demands. This can result in multiple crashes per session on high-end rigs, as reported by users transitioning from MacBooks to pro GPUs. Quantization techniques like 4-bit or 8-bit can reduce the hardware demands significantly, sometimes by 4x in memory footprint, but they come with varying impacts on output quality depending on the model architecture. The practical implication is that inference speed gains of 2-3x on new hardware can be offset by downtime from instability. When deciding, test your specific workload on both high and low end to see if the scale is worth the maintenance overhead. The key gotcha is neglecting driver updates and framework compatibility, which bites teams rushing into enterprise-grade local deployments.

Things to Try This Week

  • Try OpenAI's updated Codex for non-coding tasks like research, planning, or creating slides to explore the expanded capabilities.
  • If working with crypto, test the Oobit AI agent card for autonomous USDT spending to see how it handles real transactions.
  • Review Portkey's tools now that Palo Alto Networks has acquired it, especially if you need to add security layers to agent deployments.
  • Experiment with local inference of MiniMax M2.7 on high-memory hardware and compare uptime against lighter 16GB setups.
  • Read Simon Willison's breakdown of the Zig AI ban to decide if similar contributor-focused policies fit your open-source projects.

On the Horizon

  • Expect updates as security teams test Mythos Preview and share findings on patched vulnerabilities.
  • Watch for LM Studio support for DeepSeek v4 Flash to expand local model options.
  • Look for further details on SI's FDM1 scaling for knowledge work and computer use agents.
  • Monitor how other labs adopt early-access programs for high-capability models in sensitive domains.

Sources

Full Episode Transcript
Hey, welcome to Models and Agents, episode thirty-eight. It's May first, twenty twenty-six. Let's see what happened in the A I world today. And trust me, it's been busy. An-thropic gives defenders early access to Mythos Preview to patch A I cyber vulnerabilities before wider release. An-thropic introduced Mythos Preview as a major step up in cyber capabilities tied to coding proficiency. The company is limiting initial access to security teams for vulnerability patching. A parallel study of one million Claude conversations shaped training changes that lowered sycophancy rates in Opus four point seven and Mythos Preview for personal guidance tasks. Watch how labs balance rapid capability gains with controlled rollouts this week. An-thropic has released Mythos Preview, representing a particularly large step up in cyber capabilities that arise as part of general proficiency at coding. Rather than making it available to everyone immediately, the company is giving defenders early controlled access to find and patch vulnerabilities before Mythos-class models proliferate. This approach is paired with insights from studying one million Claude conversations. About six percent of those were requests for personal guidance primarily in health, career, relationships, and personal finance. Sycophancy was limited to nine percent overall but higher in spirituality and relationship topics. Targeted training reduced it substantially in Opus four point seven and even more in Mythos Preview. Developers working on agentic systems or secure A I deployments should pay close attention to these controlled rollout strategies. Watch for how this early access influences the broader ecosystem as similar high-capability models emerge from other labs. Open A I announced a significant upgrade to Codex, making it suitable for non-coding computer work in addition to its traditional coding tasks. The tool now assists with research, planning, documentation, slides, spreadsheets, and more. Suggested prompts are available for users to explore daily. This expansion allows knowledge workers to leverage A I for a wider array of productivity tasks without switching tools. Andrej Kar-pathy expressed excitement about SI scaling the FDM1 concept for knowledge work and computer use. He sees it as a promising direction for agent development. He referenced ongoing discussions around L L M jaggedness tied to economics and RL, as well as the emerging agent-native economy and agentic engineering practices. This highlights community interest in advancing specialized agent frameworks beyond current general models. Oobit, backed by Tether, has unveiled an A I agent card enabling autonomous spending of USDT. This tool allows for automated crypto transactions without manual intervention. Developers in the crypto space can explore this for building autonomous financial agents. Palo Alto Networks has acquired Portkey, a move aimed at securing A I agents. This acquisition enhances capabilities in protecting agentic systems from threats. Teams building A I agents should look into the integrated security features post-acquisition. Simon Willison explains why the Zig project's blanket ban on A I assisted code contributions makes sense. The focus is on the value of P R reviews for growing new contributors rather than just code quality. The decision is not based on L L M code being inferior but on long-term project health. Builders should consider similar community policies when managing open-source projects. A Reddit user in the Local Lah-mah community detailed their progression from M three Ultra to high-end setups like RTX Pro six thousand and five hundred twelve GB Mac Studios. Their current favorite is Mini Max M two point seven two hundred thirty B slash A ten B. They note that the sixteen GB MacBook Pro was surprisingly more stable than larger setups. This highlights the practical challenges and excitement in running large models locally. Speaking of running these models locally, everyone assumes bigger hardware equals smoother local A I runs. But the engineering reality involves balancing raw power against system fragility. At the base level, adding VRAM allows loading larger models without offloading to slower storage. Yet this demands precise alignment with the model's architecture and quantization scheme. Moving to two hundred fifty six GB or five hundred twelve GB configurations opens up two hundred B plus parameter models. But it often exposes bugs in memory allocation that smaller sixteen GB setups avoid due to simpler resource demands. This can result in multiple crashes per session on high-end rigs, as reported by users transitioning from MacBooks to pro G P U's. Quantization techniques like four bit or eight bit can reduce the hardware demands significantly, sometimes by four times in memory footprint. But they come with varying impacts on output quality depending on the model architecture. The practical implication is that inference speed gains of two to three times on new hardware can be offset by downtime from instability. When deciding, test your specific workload on both high and low end to see if the scale is worth the maintenance overhead. The key gotcha is neglecting driver updates and framework compatibility, which bites teams rushing into enterprise-grade local deployments. Try Open A I's updated Codex for non-coding tasks like research, planning, or creating slides to explore the expanded capabilities. If you are working with crypto, test the Oobit A I agent card for autonomous USDT spending to see how it handles real transactions. Review Portkey's tools now that Palo Alto Networks has acquired it, especially if you need to add security layers to agent deployments. Experiment with local inference of Mini Max M two point seven on high-memory hardware and compare uptime against lighter sixteen GB setups. Read Simon Willison's breakdown of the Zig A I ban to decide if similar contributor-focused policies fit your open-source projects. Expect updates as security teams test Mythos Preview and share findings on patched vulnerabilities. Watch for LM Studio support for Deep Seek v four Flash to expand local model options. Look for further details on SI's FDM1 scaling for knowledge work and computer use agents. Monitor how other labs adopt early-access programs for high-capability models in sensitive domains. Tomorrow, keep an eye on LM Studio support for Deep Seek v four Flash. That's Models and Agents for today. If you found this useful, share it with someone who's trying to keep up with all these changes, and subscribe so you don't miss tomorrow's update. The A I world moves fast. We'll help you keep up. See you tomorrow. This podcast is curated by Patrick but generated using AI voice synthesis of my voice using ElevenLabs. The primary reason to do this is I unfortunately don't have the time to be consistent with generating all the content and wanted to focus on creating consistent and regular episodes for all the themes that I enjoy and I hope others do as well.

Enjoy this episode? Get Models & Agents in your inbox

New episode alerts — no spam, unsubscribe anytime.