Player Home
All Shows
Models & Agents Planetterrian Daily Omni View Models & Agents for Beginners Fascinating Frontiers Modern Investing Techniques Tesla Shorts Time Environmental Intelligence Финансы Просто Привет, Русский!
Blogs
All Blog Posts Models & Agents Blog Planetterrian Daily Blog Omni View Blog Models & Agents for Beginners Blog Fascinating Frontiers Blog Modern Investing Techniques Blog Tesla Shorts Time Blog Environmental Intelligence Blog Финансы Просто Blog Привет, Русский! Blog
Models & Agents Models & Agents Blog

Models & Agents — Episode 24

What You Need to Know: Hugging Face just shipped TRL v1.0, turning the messy post-training pipeline (SFT → Reward Modeling → DPO/GRPO) into a stable, production-ready unified API. Liquid AI dropped LF...

April 01, 2026 Ep 24 6 min read Listen to podcast View summaries

What You Need to Know: Hugging Face just shipped TRL v1.0, turning the messy post-training pipeline (SFT → Reward Modeling → DPO/GRPO) into a stable, production-ready unified API. Liquid AI dropped LFM2.5-350M, a 350M-parameter model trained on 28T tokens with scaled RL that challenges the “bigger is better” scaling narrative. A wave of strong arXiv papers today also delivers new techniques for continual pre-training, structured reasoning, memory architectures, and quantization.

Top Story

Hugging Face has officially released TRL (Transformer Reinforcement Learning) v1.0, transitioning the library from a research-oriented codebase to a stable, production-ready framework.

The new version codifies the entire post-training pipeline—Supervised Fine-Tuning (SFT), Reward Modeling, and Alignment (DPO and the newly supported GRPO)—into a single, standardized API. This unification dramatically reduces the friction developers face when moving from research experiments to reliable production alignment workflows.

For practitioners, this means you can now maintain one coherent stack instead of juggling separate scripts and versioning headaches across the alignment stages. Teams building fine-tuned models or preference-optimized agents should care deeply—this is the kind of infrastructure improvement that saves weeks of engineering time.

What to try: Install the new TRL v1.0 and migrate an existing DPO or SFT script to the unified trainer API. Watch for GRPO examples as this reinforcement learning variant gains traction.

Source: marktechpost.com

Model Updates

Liquid AI Releases LFM2.5-350M: Intelligence Density Over Parameter Count

Liquid AI released LFM2.5-350M, a compact 350M parameter model trained on 28T tokens (up from 10T) using additional pre-training and large-scale reinforcement learning. Rather than chasing raw scale, the model serves as a case study in “intelligence density,” showing that careful data scaling and RL can extract surprising capability from smaller architectures. This is worth watching for teams deploying on-device or latency-sensitive applications where parameter count directly impacts cost and speed.

Source: marktechpost.com

PolarQuant: Near-Lossless LLM Weight Quantization via Hadamard Rotation

Researchers introduced PolarQuant, a post-training quantization method that uses block-wise normalization, Walsh-Hadamard rotation, and Gaussian-matched centroids. Hadamard rotation alone delivers 98% of the quality gain, reducing Qwen3.5-9B perplexity from 6.90 (absmax Q5) to 6.40—only +0.03 above FP16. It also serves as an excellent preprocessing step before INT4 quantization with torchao. Practical takeaway: you can now quantize aggressively with almost no quality loss.

Source: arxiv.org

OptiMer: Post-Hoc Optimal Data Composition for Continual Pre-Training

A new method called OptiMer decouples data mixture ratios from training by training one CPT model per dataset, extracting distribution vectors, then using Bayesian optimization to find ideal composition weights after the fact. On Gemma 3 27B it outperforms traditional data mixing with 15-35× lower search cost across languages and domains. The same vector pool can be re-optimized for new objectives without retraining.

Source: arxiv.org

Agent & Tool Developments

APEX-EM: Non-Parametric Online Learning for Autonomous Agents

APEX-EM introduces a structured procedural-episodic experience replay system that lets LLM agents accumulate, retrieve, and reuse plans without updating weights. Using a Plan-Retrieve-Generate-Iterate-Ingest workflow and hybrid retrieval (semantic + structural DAG matching), it delivers massive gains: +48.3pp on KGQAGen-10k (89.6% vs 41.3%) and +29.4pp on BigCodeBench. This is one of the strongest demonstrations yet that explicit memory architectures can dramatically improve agent reliability.

Source: arxiv.org

Human-Like Lifelong Memory Architecture

A new neuroscience-grounded memory framework based on complementary learning systems, valence vectors, belief hierarchies, and dual-process cognition aims to solve the “no persistent structured memory” problem in LLMs. The architecture defaults to fast System-1 retrieval with System-2 escalation and converges toward cheaper inference as experience grows. While still theoretical, the seven functional properties provide a concrete blueprint for building truly lifelong agents.

Source: arxiv.org

Practical & Community

CrossTrace: Grounded Scientific Reasoning Traces for Hypothesis Generation

CrossTrace provides 1,389 cross-domain (biomedical, AI/ML, cross-domain) grounded reasoning traces with step-level verification. Fine-tuning Qwen2.5-7B-Instruct on this data via QLoRA dramatically improves hypothesis generation quality and structural compliance. Balanced cross-domain training outperforms single-domain, suggesting scientific reasoning patterns transfer. Excellent resource for anyone building research agents or scientific copilots.

Source: arxiv.org

LiteCoST: Long-Document QA with Chain-of-Structured-Thought and SLMs

LiteCoST uses a two-pillar approach—Chain-of-Structured-Thought (CoST) traces generated by a strong LLM, then distilled into 3B/7B SLMs via SFT + GRPO. It achieves LLM-comparable quality on multi-domain long-document QA while delivering 2-4× lower latency than GPT-4o or DeepSeek-R1. The GitHub repo is already available for teams tired of throwing massive models at noisy documents.

Source: arxiv.org

MemRerank: Preference Memory for Personalized Product Reranking

MemRerank distills long purchase histories into concise, query-independent preference signals for LLM shopping agents. Using reinforcement learning to train the memory extractor against downstream reranking performance, it improves 1-in-5 selection accuracy by up to +10.61 points over raw-history and no-memory baselines. Strong evidence that explicit preference memory beats dumping history into context.

Source: arxiv.org

Under the Hood: Heuristic Override in LLM Reasoning

Everyone talks about LLMs “reasoning” as if they perform logical inference like a human. In practice, they are heavily driven by surface heuristics that can completely override implicit feasibility constraints.

The core insight is that certain salient cues (distance, cost, efficiency, semantic similarity) create strong sigmoid-shaped influences on model output. Token-level attribution shows these decisions look more like keyword associations than compositional reasoning. One study found the distance cue exerts 8.7 to 38 times more influence than the actual goal in the “car wash problem.”

The Heuristic Override Benchmark (HOB) reveals this is widespread: under strict evaluation no model exceeds 75% accuracy across 14 tested models, with presence constraints being particularly difficult (44%). A minimal hint emphasizing the key object recovers +15 percentage points on average, indicating the failure is in constraint inference rather than missing knowledge.

This adds another data point to the growing realization that current post-training heavily rewards surface patterns that correlate with correct answers during training, even when those patterns conflict with deeper logical structure.

The practical engineering guidance: when building reasoning pipelines, always include explicit precondition enumeration steps (goal-decomposition prompting recovers +6–9pp). For high-stakes domains, consider hybrid systems where a smaller verifier model checks for heuristic override patterns before accepting an answer. The gotcha that bites most teams is assuming that high benchmark scores on standard tasks mean the model is actually reasoning rather than pattern-matching.

Things to Try This Week

  • Try the new TRL v1.0 unified API for your next DPO or GRPO project — the standardized pipeline will save you significant integration time.
  • Experiment with PolarQuant on your 7B–13B models — the Hadamard rotation trick gives near-lossless Q5 quantization with almost no calibration data needed.
  • Test LiteCoST on your long-document QA tasks if you want GPT-4o quality at 2-4× lower latency using 7B SLMs.
  • Play with APEX-EM style structured memory on repetitive agent tasks — the +29–48pp gains on benchmarks suggest real productivity improvements are possible today.
  • Fine-tune a small model on CrossTrace if you’re building any kind of scientific or hypothesis-generation agent.

On the Horizon

  • More production-grade GRPO implementations and examples expected to follow the TRL v1.0 release.
  • Continued research into non-parametric memory systems as the APEX-EM and lifelong memory papers suggest this direction is gaining momentum.
  • Expect further work on heuristic-override mitigation techniques as the HOB benchmark provides a clear measuring stick.
  • Watch for additional intelligence-density focused small models as Liquid AI’s approach challenges conventional scaling wisdom.

Sources