Start Here How to Listen About Player Home
All Shows
Models & Agents Planetterrian Daily Omni View Models & Agents for Beginners Fascinating Frontiers Modern Investing Techniques Tesla Shorts Time Environmental Intelligence Финансы Просто Привет, Русский!
Blogs
All Blog Posts Models & Agents Blog Planetterrian Daily Blog Omni View Blog Models & Agents for Beginners Blog Fascinating Frontiers Blog Modern Investing Techniques Blog Tesla Shorts Time Blog Environmental Intelligence Blog Финансы Просто Blog Привет, Русский! Blog
Models & Agents Models & Agents Blog

Qwen3.6-27B paired with llama.cpp speculative decoding delivers 10x token speedu... — Episode 34

Qwen3.6-27B paired with llama.cpp speculative decoding delivers 10x token speedups in real coding sessions, hitting 136 t/s on consumer hardware.

April 23, 2026 Ep 34 6 min read Listen to podcast View summaries

Qwen3.6-27B paired with llama.cpp speculative decoding delivers 10x token speedups in real coding sessions, hitting 136 t/s on consumer hardware.

What You Need to Know: The standout story today is the dramatic inference acceleration developers are seeing with the new Qwen3.6-27B model when using ngram-based speculative decoding in llama.cpp. Community members report generation speeds climbing from 13 t/s to over 136 t/s within the same session on dual-GPU consumer rigs, while the model shows strong coding judgment, bug fixing from screenshots, and aesthetic code output that rivals much larger models. Several arXiv papers also dropped exploring hallucination neurons, stereotype localization, memory architectures for agents, and KV-cache optimizations. Pay attention to the rapid maturation of local inference tooling—this week’s practical gains feel more impactful than many benchmark headlines.

Top Story

Google Cloud AI Research and UIUC introduced ReasoningBank, a memory framework that lets LLM agents distill generalizable reasoning strategies from both their successes and failures.

The system combines experience distillation with test-time scaling, allowing agents to incrementally improve their reasoning policies rather than treating each interaction as stateless. Unlike traditional replay buffers that simply store trajectories, ReasoningBank extracts reusable strategies, creating a form of procedural memory that transfers across tasks.

Early results suggest agents genuinely get better over time instead of plateauing, which has been a persistent weakness in production agent deployments. Developers building autonomous workflows should watch this closely—persistent reasoning improvement is one of the missing pieces for agents handling core business operations.

The work directly addresses the “trust” question many enterprises are asking before they let agents touch real processes. Expect follow-up releases and open-source implementations as the team refines the approach.

Source: marktechpost.com

Model Updates

Qwen3.6-27B shows strong coding capability: r/LocalLLaMA

Developers switching from OpenAI APIs to Qwen3.6-27B in OpenCode for Svelte 5 development report near-perfect results on first try, even when paid APIs were failing. The 27B model delivers production-grade code quality that many users say exceeds expectations for its size, particularly in frontend frameworks. This continues Alibaba’s strong momentum in open coding models and suggests the next 12 months of local coding agents will be highly competitive.

Source: reddit.com

Qwen-3.6-27B with speculative decoding delivers massive speedups: r/LocalLLaMA

One developer’s session showed token generation climbing from 13.6 t/s to 136.75 t/s on the same Qwen3.6-27B-Q8_0 model simply by adding --spec-type ngram-mod with appropriate ngram and draft settings in llama-server. The model successfully debugged browser console errors from screenshots and produced high-quality, aesthetic code throughout. With 40 GB VRAM and 128 GB RAM, this setup makes 27B-class models feel remarkably responsive for iterative coding workflows.

Do Hallucination Neurons Generalize? arXiv

New research across six domains and five open-weight models (3B–8B) finds that “hallucination neurons” identified in one domain transfer poorly to others (AUROC drops from 0.783 within-domain to 0.563 cross-domain). This indicates hallucination mechanisms are largely domain-specific rather than a universal sparse signature. The finding has immediate implications for anyone building neuron-level detectors—they must be calibrated per domain rather than trained once and applied broadly.

Source: arxiv.org

Can We Locate and Prevent Stereotypes in LLMs? arXiv

Researchers mapped stereotype-related activations in GPT-2 Small and Llama 3.2, identifying both individual contrastive neurons and attention heads responsible for biased outputs. The work provides initial “bias fingerprints” that could enable targeted mitigation. While early stage, it moves the field from behavioral observation toward mechanistic understanding of where stereotypes actually live inside transformer weights.

Agent & Tool Developments

Cognis: Context-Aware Memory for Conversational AI Agents arXiv

Lyzr Cognis offers a unified memory architecture using dual-store retrieval (BM25 + Matryoshka embeddings fused by Reciprocal Rank Fusion), context-aware ingestion that checks existing memories before writing, temporal boosting, and a BGE-2 reranker. It achieves state-of-the-art results on LoCoMo and LongMemEval benchmarks and is already deployed in production. The open-source release gives developers a practical path to persistent, personalized conversational agents without starting from scratch.

OThink-SRR1: Search, Refine and Reasoning with Reinforced Learning arXiv

This framework tackles noisy retrieval and high latency in multi-hop RAG by introducing an iterative Search-Refine-Reason loop trained with GRPO-IR reinforcement learning. The Refine stage distills documents into concise facts before reasoning, and the reward model penalizes excessive retrievals. It outperforms strong baselines on four multi-hop QA benchmarks while using fewer steps and tokens, making it a promising base for information-seeking agents.

Practical & Community

We benchmarked 18 LLMs on OCR (7k+ calls): r/MachineLearning

A team ran 7,560 identical OCR/document extraction calls across 18 models on 42 standard documents and found that smaller, older, cheaper models frequently match or beat flagship models on accuracy while costing far less. They open-sourced the full dataset, framework, leaderboard, and a free tool for testing your own documents at https://github.com/ArbitrHq/ocr-mini-bench. If your workflows default to the newest model for OCR, this benchmark is worth auditing against.

Source: reddit.com

2b or not 2b? Custom LLM Scheduling Competition: r/MachineLearning

A new Kaggle competition tasks participants with deciding when to run a small 2B model versus skipping inference entirely on MMLU-style questions, optimizing a cost-based metric that penalizes both unnecessary compute and missed correct answers. The setup is deliberately simple to encourage creative classifiers or routing rules. It’s a practical first step toward intelligent routing systems that could dramatically cut token costs in production.

Source: reddit.com

Under the Hood: Temporal-Tiered KV Cache Design

Everyone talks about KV cache as if it’s just a big lookup table you grow linearly with context. In practice, TTKV shows it can be re-architected like a human memory system with fast recent memory and slower archival layers.

The core insight is that not all tokens are equally important: recent tokens need low-latency, high-precision access while older ones can tolerate lower precision and higher access cost. TTKV therefore partitions the cache into temporal tiers—recent blocks stay in fast HBM at full precision, older blocks migrate to DRAM at reduced precision—mirroring how we keep immediate context crystal clear and distant context fuzzier.

Block-wise streaming attention overlaps communication and computation when pulling from slow tiers, hiding much of the latency penalty. The paper reports 5.94× reduction in cross-tier traffic on 128K-context tasks, translating to up to 76% lower latency and 2× higher throughput versus strong baselines.

The quality gain comes with real engineering tradeoffs: you need careful tier sizing, precision scheduling logic, and migration policies that don’t thrash. Get the layout wrong and you lose the latency wins.

When to use this versus uniform KV cache? If you’re regularly running 32K+ context workloads on hardware with heterogeneous memory (HBM + DRAM or even CPU offload), TTKV-style tiering is becoming essential. The gotcha that bites most teams is assuming uniform importance across the entire context window—temporal locality is stronger than most attention visualizations suggest.

Things to Try This Week

  • Try Qwen3.6-27B-Q8_0.gguf with --spec-type ngram-mod --spec-ngram-size-n 24 in llama.cpp for your next coding session—watch the speed climb dramatically as the draft model warms up.
  • Test the ArbitrHq OCR mini-bench on your own document set before defaulting to the latest flagship model; you may cut costs by 5-10× with zero accuracy loss.
  • Experiment with Cognis memory components if you’re building conversational agents—swap in the dual-store retriever and context-aware ingestion to get persistent personalization without heavy infrastructure.
  • Enter the “2b or not 2b” Kaggle competition even if you only submit a simple classifier; the routing mindset it forces is directly transferable to production cost optimization.
  • Read the ReasoningBank paper and consider how you might log agent successes/failures in your own workflows—this distillation approach could be the next big unlock for reliable agents.

On the Horizon

  • More domain-specific fine-tunes of Qwen3.6 expected in the coming weeks as the community digests its coding and multimodal strengths.
  • Follow-up releases on ReasoningBank likely as Google Cloud AI Research turns the memory framework into a reusable library.
  • Continued progress on neuron-level interpretability—expect papers that move from locating stereotypes and hallucinations toward actual editing techniques.
  • Watch for consumer inference hardware announcements; the Reddit discussion on “Llama in a box” reflects growing frustration that could finally push vendors to ship dedicated edge chips this year.

Sources

Full Episode Transcript
Hey, welcome to Models and Agents, episode thirty-four, for April twenty-third, twenty twenty-six. Your daily A I briefing. Let's see what happened in the A I world today. And trust me, it's been busy. The standout story today is the dramatic inference acceleration developers are seeing with the new Qwenthree point six to twenty-sevenB model when using ngram-based speculative decoding in Lah-mah.cpp. Community members report generation speeds climbing from thirteen tokens per second to over one hundred thirty six tokens per second within the same session on dual G P U consumer rigs. At the same time the model shows strong coding judgment, bug fixing from screenshots, and aesthetic code output that rivals much larger models. Several arXiv papers also dropped exploring hallucination neurons, stereotype localization, memory architectures for agents, and KV cache optimizations. Pay attention to the rapid maturation of local inference tooling. This week’s practical gains feel more impactful than many benchmark headlines. What you need to know is that local models just got a whole lot more usable for real coding work, and agent memory research is finally moving beyond stateless resets. The top story today comes from Google Cloud A I Research and UIUC. They introduced ReasoningBank, a memory framework that lets L L M agents distill generalizable reasoning strategies from both their successes and failures. The system combines experience distillation with test time scaling. This allows agents to incrementally improve their reasoning policies rather than treating each interaction as stateless. Unlike traditional replay buffers that simply store trajectories, ReasoningBank extracts reusable strategies. It creates a form of procedural memory that transfers across tasks. Early results suggest agents genuinely get better over time instead of plateauing, which has been a persistent weakness in production agent deployments. Developers building autonomous workflows should watch this closely. Persistent reasoning improvement is one of the missing pieces for agents handling core business operations. The work directly addresses the trust question many enterprises are asking before they let agents touch real processes. Expect follow up releases and open source implementations as the team refines the approach. Now on the model side, Qwenthree point six to twenty-sevenB is turning heads in the local L L M community. Developers switching from Open A I A P I's to Qwenthree point six to twenty-sevenB in OpenCode for Svelte five development report near perfect results on first try. Even when the paid A P I's were failing, this twenty seven billion parameter model delivered production grade code quality. Many users say it exceeds expectations for its size, particularly in frontend frameworks. This continues Alibaba’s strong momentum in open coding models. It suggests the next twelve months of local coding agents will be highly competitive. The real fireworks though came when one developer paired Qwenthree point six to twenty-sevenB with speculative decoding in Lah-mah.cpp. Using the same Qwenthree point six to twenty-sevenB-Q8_0 model, token generation climbed from thirteen point six tokens per second to one hundred thirty six point seven five tokens per second. All it took was adding the ngram-mod speculative decoding flag with appropriate settings in Lah-mah server. The model successfully debugged browser console errors from screenshots. It produced high quality, aesthetic code throughout the entire session. With forty gigabytes of VRAM and one hundred twenty eight gigabytes of RAM, this setup makes twenty seven billion parameter class models feel remarkably responsive for iterative coding workflows. That is the kind of leap that changes how you actually use these models day to day. On the research front, a new arXiv paper asks whether hallucination neurons generalize. The study looked across six domains and five open weight models ranging from three to eight billion parameters. It found that hallucination neurons identified in one domain transfer poorly to others. The area under the receiver operating characteristic curve drops from zero point seven eight three within domain to zero point five six three cross domain. This indicates hallucination mechanisms are largely domain specific rather than a universal sparse signature. The finding has immediate implications for anyone building neuron level detectors. They must be calibrated per domain rather than trained once and applied broadly. Another paper explores whether we can locate and prevent stereotypes in large language models. Researchers mapped stereotype related activations in both G P T two Small and Lah-mah three point two. They identified individual contrastive neurons and attention heads responsible for biased outputs. The work provides initial bias fingerprints that could enable targeted mitigation. While still early stage, it moves the field from behavioral observation toward mechanistic understanding of where stereotypes actually live inside transformer weights. Shifting to agent and tool developments, Lyzr released Cognis, an open source context aware memory architecture for conversational A I agents. It uses dual store retrieval combining BM twenty five with Matryoshka embeddings fused by reciprocal rank fusion. The system adds context aware ingestion that checks existing memories before writing new ones, plus temporal boosting and a BGE two reranker. Cognis achieves state of the art results on the LoCoMo and LongMemEval benchmarks. It is already deployed in production, which gives developers a practical path to persistent, personalized conversational agents without starting from scratch. Another interesting framework is OThink-SRR1, which tackles noisy retrieval and high latency in multi hop retrieval augmented generation. It introduces an iterative search, refine, and reason loop trained with GRPO-IR reinforcement learning. The refine stage distills documents into concise facts before reasoning begins. A reward model penalizes excessive retrievals to keep things efficient. The approach outperforms strong baselines on four multi hop question answering benchmarks while using fewer steps and tokens. It looks like a promising base for information seeking agents. In the practical and community space, a team benchmarked eighteen different large language models on optical character recognition and document extraction. They ran seven thousand five hundred sixty identical calls across forty two standard documents. The surprising result is that smaller, older, cheaper models frequently match or beat flagship models on accuracy while costing far less. The team open sourced the full dataset, framework, leaderboard, and a free tool for testing your own documents. If your workflows default to the newest model for OCR tasks, this benchmark is worth auditing against. There is also a new Kaggle competition called two B or not two B. It tasks participants with deciding when to run a small two billion parameter model versus skipping inference entirely on M M L U style questions. The goal is optimizing a cost based metric that penalizes both unnecessary compute and missed correct answers. The setup is deliberately simple to encourage creative classifiers or routing rules. It is a practical first step toward intelligent routing systems that could dramatically cut token costs in production. OK, let us pop the hood on this temporal tiered KV cache design that has been generating discussion. Everyone talks about the KV cache as if it is just a big lookup table you grow linearly with context. In practice, this new approach shows it can be re-architected like a human memory system with fast recent memory and slower archival layers. The core insight is that not all tokens are equally important. Recent tokens need low latency, high precision access. Older ones can tolerate lower precision and higher access cost. The temporal tiered KV cache therefore partitions the cache into temporal tiers. Recent blocks stay in fast high bandwidth memory at full precision. Older blocks migrate to dynamic random access memory at reduced precision. This mirrors how we keep immediate context crystal clear and distant context fuzzier. Block wise streaming attention overlaps communication and computation when pulling from slow tiers. That hides much of the latency penalty. The paper reports a five point nine four times reduction in cross tier traffic on one hundred twenty eight thousand context tasks. That translates to up to seventy six percent lower latency and two times higher throughput versus strong baselines. The quality gain comes with real engineering tradeoffs. You need careful tier sizing, precision scheduling logic, and migration policies that do not thrash. Get the layout wrong and you lose the latency wins. So when should you actually reach for this versus uniform KV cache. If you are regularly running thirty two thousand plus context workloads on hardware with heterogeneous memory, whether high bandwidth memory plus dynamic random access memory or even CPU offload, this tiering approach is becoming essential. The gotcha that bites most teams is assuming uniform importance across the entire context window. Temporal locality is stronger than most attention visualizations suggest. If you have not tried Qwenthree point six to twenty-sevenB-Q8_0 dot GGUF with the ngram speculative decoding flags in Lah-mah dot cpp, this week is the perfect time. Watch the speed climb dramatically as the draft model warms up during your next coding session. Test the ArbitrHq OCR mini bench on your own document set before defaulting to the latest flagship model. You may cut costs by five to ten times with zero accuracy loss. Experiment with Cognis memory components if you are building conversational agents. Swap in the dual store retriever and context aware ingestion to get persistent personalization without heavy infrastructure. Enter the two B or not two B Kaggle competition even if you only submit a simple classifier. The routing mindset it forces is directly transferable to production cost optimization. Read the ReasoningBank paper and consider how you might log agent successes and failures in your own workflows. This distillation approach could be the next big unlock for reliable agents. On the horizon, more domain specific fine tunes of Qwen3.6 are expected in the coming weeks as the community digests its coding and multimodal strengths. Follow up releases on ReasoningBank are likely as Google Cloud A I Research turns the memory framework into a reusable library. Continued progress on neuron level interpretability should bring papers that move from locating stereotypes and hallucinations toward actual editing techniques. Keep an eye on consumer inference hardware announcements as the growing frustration around local performance could finally push vendors to ship dedicated edge chips this year. Before we go, tomorrow keep an eye on whether any early open source implementations of ReasoningBank appear. That is Models and Agents for today. If you found this useful, share it with someone who is trying to keep up with all these changes, and subscribe so you do not miss tomorrow's update. The A I world moves fast. We will help you keep up. See you tomorrow. This podcast is curated by Patrick but generated using AI voice synthesis of my voice using ElevenLabs. The primary reason to do this is I unfortunately don't have the time to be consistent with generating all the content and wanted to focus on creating consistent and regular episodes for all the themes that I enjoy and I hope others do as well.

Enjoy this episode? Get Models & Agents in your inbox

New episode alerts — no spam, unsubscribe anytime.