Start Here How to Listen About Player Home
All Shows
Models & Agents Planetterrian Daily Omni View Models & Agents for Beginners Fascinating Frontiers Modern Investing Techniques Tesla Shorts Time Environmental Intelligence Финансы Просто Привет, Русский!
Blogs
All Blog Posts Models & Agents Blog Planetterrian Daily Blog Omni View Blog Models & Agents for Beginners Blog Fascinating Frontiers Blog Modern Investing Techniques Blog Tesla Shorts Time Blog Environmental Intelligence Blog Финансы Просто Blog Привет, Русский! Blog
Models & Agents Models & Agents Blog

Models & Agents — Episode 39

> Mistral AI launches a 128B model with remote agents and strong coding performance.

May 03, 2026 Ep 39 5 min read Listen to podcast View summaries

Models & Agents

Mistral AI launches a 128B model with remote agents and strong coding performance.

What You Need to Know: Mistral AI released Mistral Medium 3.5 alongside remote agents in Vibe for async cloud coding sessions and an agentic Work mode in Le Chat. The launch focuses on practical developer tools for building more autonomous coding workflows. Watch how these capabilities integrate into existing agent stacks this week.


Top Story

Mistral AI has released Mistral Medium 3.5, a new 128B flagship model, together with Remote Agents in Vibe that support async cloud-based coding sessions. The model includes an agentic Work mode in Le Chat designed for structured developer interactions. This combination targets teams building AI agents that need reliable coding assistance without constant human oversight. Developers can now run longer autonomous sessions on cloud infrastructure while maintaining control through the platform's interface. The release marks a step toward more production-ready agent tooling from Mistral. Expect community experiments with the remote agent features to surface practical integration patterns soon. Source: marktechpost.com


Model Updates

DeepSeek V4 Leads Chinese Models in New Evaluation

The CAISI report identifies DeepSeek V4 as the strongest model currently available in China. It still trails leading US systems by roughly eight months according to the evaluation. The findings provide a snapshot of regional progress in open model development. Source: reddit.com

Qwen3.6-27B and Coder-Next Show Closely Matched Results

Extensive side-by-side testing on high-end GPUs found the two models perform similarly across many tasks. Coder-Next delivered better cost efficiency on bounded document tasks while the 27B variant handled certain research-style queries more effectively. The comparison highlights how different architectural choices trade off consistency against efficiency. Source: reddit.com

Qwen 3.6 and Gemma 4 Reveal Distinct Vision Strengths

Local tests on M1 Max hardware showed Gemma 4 following formatting instructions more reliably for tasks like bounding boxes and Western cultural content. Qwen 3.6 performed better on video tracking and Asian context recognition but required strict 2 FPS preprocessing for video inputs. The results underscore how training data geography affects real-world behavior. Source: reddit.com

Ten Local Image Generation Models Compared on Mac

Tests across models from SD 1.5 through Flux dev, Qwen-Image, and Gemini evaluated photorealism, text rendering, and cultural accuracy on M1 Max hardware. Qwen-Image Lightning offered a strong speed-quality balance while Flux led in photorealism. Cultural biases appeared more tied to training data origins than model scale. Source: reddit.com


Agent & Tool Developments

Multi-Agent AI Workflow for Biological Network Modeling

The guide walks through constructing a multi-agent system that coordinates tasks across protein interactions, metabolism, and cell signaling simulations. Agents handle specialized sub-problems in a modular way that supports complex scientific workflows. Developers working in bioinformatics can adapt the pattern for other domain-specific orchestration needs. Source: marktechpost.com

FPGA Approaches for Speculative Decoding

Community discussion examines building FPGAs for models in the 20-30M parameter range with quantization and pairing them with speculative decoding using fast smaller models. The approach aims to deliver high token throughput at lower hardware cost. Questions remain around scaling beyond current limits and integration with existing inference stacks. Source: reddit.com

Fast Memory Mechanism on Frozen Small Transformers

A toy experiment demonstrates how a frozen Pythia-70M model can use forward-derived correction vectors for one-shot symbolic recall without any weight updates. The setup separates conflicting context meanings through learned retrieval geometry. It offers a lightweight path toward in-context adaptation that avoids full fine-tuning. Source: reddit.com


Practical & Community

Complete Transformer Built in Pure C++17

A developer implemented a full GPT-style decoder-only transformer from scratch using only the C++17 standard library, including hand-written tensor operations, attention, and analytical backpropagation. The project trains on CPU with no external dependencies and includes an OpenMP-accelerated version. The repo provides a clear reference for understanding every layer of a transformer without framework abstractions. Source: reddit.com

SAE Fine-Tune of Qwen 3.5 Released on Hugging Face

The Qwen/SAE-Res-Qwen3.5-27B-W80K-L0_100 model provides a ready-to-use sparse autoencoder variant for vector-based steering experiments. It opens immediate access to mechanistic interpretability techniques on a 27B-scale model. Users can load it directly for research into activation steering and feature extraction. Source: reddit.com

Python Agent Connecting Qwen to LM Studio

A user built a Python agent via Claude that links Qwen 3.6 35B running in LM Studio to generate structured output for a 2025 tax return form. The agent reads input fields and produces a template without stopping until completion. It demonstrates quick agent scaffolding for domain-specific document generation tasks.


Under the Hood: Tokenization Drift

Everyone talks about model outputs as if they flow directly from weights and prompt semantics, but the first step of turning text into token IDs often creates hidden variability. Tokenization maps strings to a fixed vocabulary of IDs, so even small formatting differences like extra spaces, line breaks, or punctuation produce different sequences. These ID changes shift the initial embedding vectors that enter the transformer, which then alters attention patterns and can send generation down entirely different paths. In RAG pipelines this shows up as inconsistent answers when the same query retrieves documents with minor formatting variations from different sources. The engineering tradeoff is that strict input normalization reduces drift but adds preprocessing latency and can occasionally remove useful signals present in the original formatting. Production teams typically insert canonical formatting early in the data pipeline to keep behavior predictable. The practical takeaway is to always test critical applications against varied input styles rather than assuming semantic equivalence guarantees identical tokenization.


Things to Try This Week

  • Test Mistral Medium 3.5 through Le Chat's Work mode on a multi-step coding task to evaluate the new remote agent capabilities.
  • Run local vision comparisons between Qwen 3.6 and Gemma 4 using vLLM on your own image or video dataset to identify which model better matches your domain.
  • Set up the multi-agent biological workflow on a small simulation problem to practice modular agent orchestration.
  • Clone the C++ transformer repo and run the CPU training example to see every component of a model implemented without frameworks.
  • Load the new SAE Qwen variant from Hugging Face and experiment with steering on a few test prompts.

On the Horizon

  • More real-world comparisons of vision models as local hardware and inference engines continue to improve.
  • Additional hardware explorations around FPGAs and speculative decoding for cost-effective local inference.
  • New fine-tuned and steering-focused releases on Hugging Face targeting popular base models.
  • Expanded guides for multi-agent systems in specialized domains like scientific computing.
Full Episode Transcript
Hey, welcome to Models and Agents, episode thirty-nine. It's May third, twenty twenty-six. Let's see what happened in the A I world today. And trust me, it's been busy. Mee-stral A I launches a one hundred twenty eight billion model with remote agents and strong coding performance. Mee-stral A I released Mee-stral Medium three point five alongside remote agents in Vibe for async cloud coding sessions. They also introduced an agentic Work mode in Le Chat designed for structured developer interactions. The launch focuses on practical developer tools for building more autonomous coding workflows. Keep an eye on how these capabilities integrate into existing agent stacks this week. The big news today comes from Mee-stral A I with their new Mee-stral Medium three point five model. This one hundred twenty eight billion parameter flagship arrives together with Remote Agents in Vibe that support async cloud based coding sessions. There is also an agentic Work mode in Le Chat for structured developer interactions. The combination targets teams building A I agents that need reliable coding assistance without constant human oversight. Developers can now run longer autonomous sessions on cloud infrastructure while maintaining control through the platform interface. The release marks a step toward more production ready agent tooling from Mee-stral. Expect community experiments with the remote agent features to surface practical integration patterns soon. Shifting to other model updates, the CAISI report identifies Deep Seek V four as the strongest model currently available in China. It still trails leading U S systems by roughly eight months according to the evaluation. The findings provide a snapshot of regional progress in open model development. Side by side testing on high end G P U's found Chwen three point six twenty seven B and Coder Next performing similarly across many tasks. Coder Next delivered better cost efficiency on bounded document tasks. The twenty seven billion variant handled certain research style queries more effectively. The comparison highlights how different architectural choices trade off consistency against efficiency. Local tests on M one Max hardware showed Gemma four following formatting instructions more reliably for tasks like bounding boxes and Western cultural content. Chwen three point six performed better on video tracking and Asian context recognition. But it required strict two frames per second preprocessing for video inputs. The results underscore how training data geography affects real world behavior. Tests across models from SD one point five through Flux dev, Chwen Image, and Gemini evaluated photorealism, text rendering, and cultural accuracy on M one Max hardware. Chwen Image Lightning offered a strong speed quality balance while Flux led in photorealism. Cultural biases appeared more tied to training data origins than model scale. On the agent side, a new guide walks through constructing a multi agent system for biological network modeling. It coordinates tasks across protein interactions, metabolism, and cell signaling simulations. Agents handle specialized sub problems in a modular way that supports complex scientific workflows. Developers working in bioinformatics can adapt the pattern for other domain specific orchestration needs. Community discussion examines building FPGAs for models in the twenty to thirty million parameter range with quantization. They pair these with speculative decoding using fast smaller models. The approach aims to deliver high token throughput at lower hardware cost. Questions remain around scaling beyond current limits and integration with existing inference stacks. A toy experiment demonstrates how a frozen Pythia seventy M model can use forward derived correction vectors for one shot symbolic recall. This happens without any weight updates. The setup separates conflicting context meanings through learned retrieval geometry. It offers a lightweight path toward in context adaptation that avoids full fine tuning. In the practical space, a developer implemented a full G P T style decoder only transformer from scratch using only the C plus plus seventeen standard library. That includes hand written tensor operations, attention, and analytical backpropagation. The project trains on CPU with no external dependencies and includes an Open MP accelerated version. The repo provides a clear reference for understanding every layer of a transformer without framework abstractions. On Hugging Face, the Chwen SAE Res Chwen three point five twenty seven B W eighty K L zero one hundred model provides a ready to use sparse autoencoder variant. It opens immediate access to mechanistic interpretability techniques on a twenty seven billion scale model. Users can load it directly for research into activation steering and feature extraction. A user built a Python agent via Claude that links Chwen three point six thirty five B running in LM Studio to generate structured output. The agent reads input fields and produces a template for a twenty twenty five tax return form without stopping until completion. It demonstrates quick agent scaffolding for domain specific document generation tasks. OK, let's pop the hood on this one and talk about tokenization drift. Everyone talks about model outputs as if they flow directly from weights and prompt semantics. But the first step of turning text into token IDs often creates hidden variability. Tokenization maps strings to a fixed vocabulary of IDs. So even small formatting differences like extra spaces, line breaks, or punctuation produce different sequences. These ID changes shift the initial embedding vectors that enter the transformer. Which then alters attention patterns and can send generation down entirely different paths. In rag pipelines this shows up as inconsistent answers when the same query retrieves documents with minor formatting variations from different sources. The engineering tradeoff is that strict input normalization reduces drift but adds preprocessing latency and can occasionally remove useful signals present in the original formatting. Production teams typically insert canonical formatting early in the data pipeline to keep behavior predictable. The practical takeaway is to always test critical applications against varied input styles rather than assuming semantic equivalence guarantees identical tokenization. If you want to get hands on this week, test Mee-stral Medium three point five through Le Chat Work mode on a multi step coding task. This will help evaluate the new remote agent capabilities. Run local vision comparisons between Chwen three point six and Gemma four using vLLM on your own image or video dataset. That will show which model better matches your domain. Set up the multi agent biological workflow on a small simulation problem to practice modular agent orchestration. Clone the C plus plus transformer repo and run the CPU training example to see every component of a model implemented without frameworks. Load the new SAE Chwen variant from Hugging Face and experiment with steering on a few test prompts. Looking ahead, expect more real world comparisons of vision models as local hardware and inference engines continue to improve. Additional hardware explorations around FPGAs and speculative decoding for cost effective local inference are likely on the way. New fine tuned and steering focused releases on Hugging Face targeting popular base models should appear soon. Expanded guides for multi agent systems in specialized domains like scientific computing are also coming. Before we go, tomorrow keep an eye on how developers are testing those FPGA approaches for speculative decoding in real inference setups. That wraps up today's A I briefing. Share this with a developer or builder who wants to stay current. Subscribe wherever you listen. See you tomorrow. This podcast is curated by Patrick but generated using AI voice synthesis of my voice using ElevenLabs. The primary reason to do this is I unfortunately don't have the time to be consistent with generating all the content and wanted to focus on creating consistent and regular episodes for all the themes that I enjoy and I hope others do as well.

Enjoy this episode? Get Models & Agents in your inbox

New episode alerts — no spam, unsubscribe anytime.