<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[LLM Watch: Deep Dives]]></title><description><![CDATA[Want to dig into the latest research without having to read dozens of pages yourself? Let me break it down for you. 

Everything I'll write about will be brand new at the time of publishing - you'll get to know about all the important bits before others have even read the abstract. ]]></description><link>https://www.llmwatch.com/s/deep-dives</link><image><url>https://substackcdn.com/image/fetch/$s_!WczK!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5d95c476-43a7-4447-9081-9298a1fc325a_1280x1280.png</url><title>LLM Watch: Deep Dives</title><link>https://www.llmwatch.com/s/deep-dives</link></image><generator>Substack</generator><lastBuildDate>Fri, 01 May 2026 00:20:35 GMT</lastBuildDate><atom:link href="https://www.llmwatch.com/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Pascal Biese]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[xaiguy@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[xaiguy@substack.com]]></itunes:email><itunes:name><![CDATA[Pascal Biese]]></itunes:name></itunes:owner><itunes:author><![CDATA[Pascal Biese]]></itunes:author><googleplay:owner><![CDATA[xaiguy@substack.com]]></googleplay:owner><googleplay:email><![CDATA[xaiguy@substack.com]]></googleplay:email><googleplay:author><![CDATA[Pascal Biese]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[Falcon-H1: The AI Chimera That Challenges The Transformer]]></title><description><![CDATA[How a new AI model beats models twice its size]]></description><link>https://www.llmwatch.com/p/falcon-h1-the-ai-chimera-that-challenges</link><guid isPermaLink="false">https://www.llmwatch.com/p/falcon-h1-the-ai-chimera-that-challenges</guid><dc:creator><![CDATA[Pascal Biese]]></dc:creator><pubDate>Sat, 02 Aug 2025 17:57:08 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/e6e93a56-8ca5-4084-9e99-e0b01eea4d9c_1536x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>Sometimes in AI, <strong>smaller</strong> can be <strong>smarter</strong>. A recent research release shows a 34-billion-parameter AI model matching the performance of models twice its size by doing something unconventional &#8211; giving the AI two different neural networks instead of one. This isn&#8217;t just a one-off trick: it hints at how the next generation of foundation models might be built, blending techniques and potentially modalities to achieve more with less. In this post, we break down this hybrid AI breakthrough and explore what it means for tech builders, researchers, and strategists alike.</em></p><h2>Meet Falcon-H1 &#8211; The Hybrid AI That Punches Above Its Weight</h2><p>A few months ago, the team at <strong>Technology Innovation Institute (TII)</strong> unveiled <strong><a href="https://github.com/tiiuae/falcon-h1">Falcon-H1</a></strong>, a series of large language models with a new hybrid architecture (and now, we finally got the <a href="https://www.arxiv.org/abs/2507.22448">report</a>). Instead of relying purely on the standard Transformer architecture that underpins most modern generative AIs, Falcon-H1 combines <strong>two different model paradigms in parallel</strong> within each layer. In simple terms, it&#8217;s like giving the model <strong>&#8220;two heads&#8221;</strong> for thinking: one head uses the classic attention mechanism (great for focusing on specific details in text), and the other uses a <strong>State Space Model (SSM)</strong> (great for remembering long sequences efficiently). These twin heads work side by side, and their outputs are fused together, allowing Falcon-H1 to leverage the strengths of both approaches.</p><p>Why go hybrid? The <strong>Transformer</strong> (attention) is excellent at understanding contextual relationships (think of it as a spotlight that highlights relevant words), while the <strong>SSM</strong> provides superior long-term memory and speed (think of it as an efficient notepad that can handle very long documents without slowing down). By marrying the two, Falcon-H1 aims to get the best of both worlds &#8211; and according to the results, it succeeds. The <strong>hybrid architecture</strong> yields <strong>faster inference and lower memory usage</strong> than a pure Transformer, without sacrificing accuracy. In fact, the researchers found that they only needed a relatively <strong>small fraction of attention heads</strong> (the Transformer part) to achieve strong performance &#8211; the SSM component carries a lot of the load for long-text understanding. This is a significant departure from the &#8220;all-Transformer, all the time&#8221; approach of most large models.</p><p><strong>So, what is Falcon-H1 exactly?</strong> It&#8217;s a family of six open-source models of varying sizes (from a tiny 0.5 billion parameters up to 34 billion) released with a permissive license (Apache 2.0). Each model comes in both a base version and an <em>instruction-tuned</em> version (trained to follow human instructions better). Despite the relatively moderate sizes (by today&#8217;s standards), Falcon-H1 models are <strong>punching above their weight class</strong>. Here are some of the highlights and features of Falcon-H1:</p><p><strong>Hybrid Architecture (Attention + SSM):</strong> Each model layer uses <em>two parallel heads</em> &#8211; one Transformer attention head and one &#8220;Mamba-2&#8221; SSM head &#8211; whose outputs are concatenated. The ratio of attention to SSM can be tuned, but it turns out you don&#8217;t need much attention to get great results. This design provides <strong>strong generalization across tasks with faster speed and better memory efficiency</strong> than attention alone. (Think of it like a car with both an electric motor and a gas engine &#8211; efficient for long hauls yet quick for bursts of power.)</p><p><strong>Wide Range of Model Sizes:</strong> Falcon-H1 is released in six sizes &#8211; <strong>0.5B, 1.5B, 1.5B-Deep, 3B, 7B, and 34B</strong> parameters &#8211; covering everything from lightweight models that could run on edge devices to hefty models for servers. The &#8220;1.5B-Deep&#8221; variant is intriguing: it has the same parameter count as the 1.5B model but <strong>far more layers (66 layers vs 24)</strong>, illustrating one of the team&#8217;s findings that <strong>deeper models can outperform wider ones</strong> under the same budget.</p><p><strong>Multilingual by Design:</strong> These models aren&#8217;t just English-centric. Falcon-H1 was trained <strong>natively on 18 languages</strong> (covering English, Chinese, Arabic, Hindi, Spanish, French, and many more) and the underlying tokenizer can scale to 100+ languages. In practice, this means the models can understand and generate multiple languages out-of-the-box, a nod to our globally connected world where AI needs to speak many tongues.</p><p><strong>Compact Models, Big Performance:</strong> Perhaps the most impressive claim: Falcon-H1 models are <strong>designed to match or exceed the performance of models at least twice their size</strong>. The <strong>tiny 0.5B model</strong> (the size of a single GPT-2, roughly) delivers performance on par with typical 7B parameter models from just a year prior. The <strong>1.5B-Deep model</strong> actually <strong>outperforms many 7B models</strong> out there. And the flagship <strong>34B model</strong> has been shown to <strong>match or beat certain 70B models</strong> (like Meta&#8217;s and Alibaba&#8217;s latest) on a range of benchmarks. In short, Falcon-H1 squeezes more juice out of every parameter &#8211; a welcome development as we confront the cost and complexity of ever-larger AI models.</p><p><strong>256K Context Length:</strong> Falcon-H1 can handle <em>really</em> long inputs. Thanks to the SSM-based design and some clever tweaking of positional embeddings, it supports up to <strong>256,000 tokens of context</strong> (for the 7B and 34B models). That&#8217;s orders of magnitude more than the 4K or 8K token limits many GPT-style models had not long ago. For perspective, 256k tokens is roughly a <strong>hundreds of pages of text</strong> in one go. This opens the door to AI that can ingest and reason about long documents (or lengthy multi-turn conversations) without breaking a sweat.</p><p><strong>Domain Expertise (STEM and Code):</strong> Because of a carefully curated training set, Falcon-H1 shows strong abilities in math, science, and coding tasks. The team deliberately included high-quality technical content &#8211; like math problem sets, scientific papers, and a large code corpus &#8211; in the training mix. They even tweaked the tokenizer to handle things like LaTeX math notation and code syntax better (e.g., splitting digits and punctuation so the model learns numbers and code structure more effectively). As a result, the models have <strong>exceptional STEM capabilities</strong> compared to peers, excelling at mathematical reasoning and scientific Q&amp;A.</p><p><strong>Open-Source and Accessible:</strong> All Falcon-H1 models are released under a permissive open-source license and are available on Hugging Face for anyone to use. Crucially, the team also provides <strong>quantized versions</strong> of the instruction-tuned models (over 30 checkpoints in total), meaning you can run these models with much less memory than the full parameter count would normally require. In practice, this could allow a 34B model to run on a single high-end GPU or even a powerful laptop &#8211; bringing near state-of-the-art AI capabilities to those without mega-scale infrastructure.</p><p>In summary, Falcon-H1 arrives not as a single monolithic model but as a <strong>portfolio of AI models</strong> that are <em>smaller, multilingual, long-winded (in a good way), and freely available</em>. It&#8217;s a bold demonstration that <strong>innovation in architecture and training can rival sheer scale</strong>. But how did the researchers actually pull this off? Let&#8217;s peek under the hood.</p><h2>How Did They Do It? Rethinking the Recipe for AI Success</h2><p>It&#8217;s one thing to list features and accomplishments, but Falcon-H1&#8217;s creation is as interesting as its results. The researchers didn&#8217;t just scale up a Transformer in the usual way &#8211; they <strong>questioned a lot of AI&#8217;s conventional wisdom</strong> and tried some decidedly off-beat ideas in the process. Here are a few of the key methods and insights behind the scenes:</p><p><strong>Hybrid &#8220;Mixer&#8221; Layers:</strong> As mentioned, the core innovation was the <strong>parallel hybrid layer design</strong>. Instead of stacking some SSM layers and some attention layers separately, Falcon-H1&#8217;s layers each contain both components running concurrently</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!B2Gm!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F504ee054-25f4-4396-891b-5660ef8c295c_613x547.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!B2Gm!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F504ee054-25f4-4396-891b-5660ef8c295c_613x547.png 424w, https://substackcdn.com/image/fetch/$s_!B2Gm!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F504ee054-25f4-4396-891b-5660ef8c295c_613x547.png 848w, https://substackcdn.com/image/fetch/$s_!B2Gm!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F504ee054-25f4-4396-891b-5660ef8c295c_613x547.png 1272w, https://substackcdn.com/image/fetch/$s_!B2Gm!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F504ee054-25f4-4396-891b-5660ef8c295c_613x547.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!B2Gm!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F504ee054-25f4-4396-891b-5660ef8c295c_613x547.png" width="613" height="547" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/504ee054-25f4-4396-891b-5660ef8c295c_613x547.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:547,&quot;width&quot;:613,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:96249,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.llmwatch.com/i/169937849?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F504ee054-25f4-4396-891b-5660ef8c295c_613x547.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!B2Gm!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F504ee054-25f4-4396-891b-5660ef8c295c_613x547.png 424w, https://substackcdn.com/image/fetch/$s_!B2Gm!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F504ee054-25f4-4396-891b-5660ef8c295c_613x547.png 848w, https://substackcdn.com/image/fetch/$s_!B2Gm!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F504ee054-25f4-4396-891b-5660ef8c295c_613x547.png 1272w, https://substackcdn.com/image/fetch/$s_!B2Gm!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F504ee054-25f4-4396-891b-5660ef8c295c_613x547.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The diagram above (from the Falcon team) illustrates this: an input goes through a normalization, then splits into an SSM branch (purple) and an Attention branch (blue). Both process the data <em>at the same time</em>, then their outputs are <strong>concatenated</strong> and passed through an MLP (feed-forward network) together before the layer outputs its result. This parallel approach, as opposed to a sequential mix, was found to work best. Crucially, it lets the designers <strong>tune the ratio</strong> of how many channels go to SSM vs Attention vs MLP. After extensive experiments, they settled on a ratio (for example, in one configuration, roughly 2 parts SSM, 1 part attention, 5 parts MLP) that optimized efficiency and learning dynamics. Interestingly, <strong>adding too much attention actually hurt performance</strong>, whereas giving the SSM a healthy share improved it. In other words, <em>more Transformer wasn&#8217;t always better</em> &#8211; a smaller attention mechanism paired with SSM was the sweet spot.</p><p><strong>Positional Encoding Trick for Long Contexts:</strong> One challenge with super-long context windows (like 256k tokens) is how to give the model a sense of position (which word comes first, second, etc.) without losing precision. Falcon-H1 tackled this by using an unusually <strong>high-frequency rotary positional embedding (RoPE)</strong> base. Essentially, they cranked up a parameter in the positional encoding formula to $10^{11}$ (far beyond typical values), which significantly improved the model&#8217;s ability to generalize on long sequences. Normally, using such a high frequency would destabilize a pure Transformer, but the hypothesis is that the <strong>SSM part naturally handles some positional information</strong>, allowing the attention part to take on this extreme frequency without trouble. It&#8217;s a bit like giving the model an extremely fine-grained ruler to measure position in text &#8211; and finding that it actually can make use of all those tick marks when reading a long document.</p><p><strong>Deeper, Not Just Wider:</strong> When allocating a fixed budget of parameters, the team found that making the model <strong>deeper (more layers)</strong> often yielded better results than making it wider (bigger layer dimensions). This insight led to the creation of the <strong>Falcon-H1-1.5B-Deep</strong> variant &#8211; it packs 66 layers into 1.5B params, compared to the standard 24 layers &#8211; and that deep model outperformed many models with 3B or even 7B parameters. It&#8217;s a notable data point in the ongoing &#8220;depth vs width&#8221; debate in model design, suggesting that for some tasks a tall, thin model can beat a short, fat one (up to a point).</p><p><strong>Unorthodox Training Curriculum:</strong> In training AI, a common practice is <em>curriculum learning</em> &#8211; feed the model easy examples first and gradually move to harder ones. Falcon-H1&#8217;s team tried the opposite and it <strong>worked better</strong>. They gave even the most <strong>complex data (like tough math problems and very long texts) right from the start of training</strong>, rather than saving them for last. Counterintuitively, this early exposure to hard examples gave the model &#8220;more time to learn&#8221; the skills needed for those challenges. It&#8217;s akin to teaching a language by mixing in advanced literature from day one alongside the basics &#8211; a risky approach for humans, but apparently effective for this AI&#8217;s development.</p><p><strong>More Data, More Reuse:</strong> Falcon-H1 was trained on a colossal <strong>20 trillion token</strong> corpus (with about 18T actually used). This includes a refined web crawl (&#8220;FineWeb&#8221;), a large multilingual set (Wikipedia, Common Crawl, subtitles, etc. for 17 languages), a 67-language code dataset (deduplicated and cleaned), and specialized math and science data. One concern with such large datasets is models simply <strong>memorizing</strong> content. The prevailing wisdom is usually to avoid reusing data in training passes. However, the team found that this risk is a bit <strong>overestimated</strong> &#8211; they carefully measured the model&#8217;s &#8220;memorization window&#8221; and concluded they could <strong>reuse high-quality examples more times</strong> than usual <strong>without hurting generalization</strong>. By doing so, they ensured that the best content (like high-quality science articles or code) had a stronger influence on training, rather than being diluted by a flood of lower-quality web text. This approach improved the model&#8217;s grasp of those domains without turning it into a rote copy-machine.</p><p><strong>Stabilizing the Training of SSMs:</strong> State-space models are powerful for long sequences, but they can be tricky &#8211; prone to training instabilities (loss spikes). The Falcon team addressed this by inserting special <strong>&#8220;dampening&#8221; factors into the SSM components</strong> as part of their parametrization (essentially, controlling the scale at which the SSM part learns). By doing so, they eliminated the nasty training spikes that often plague SSM-based networks, resulting in a smooth learning curve. They also paid close attention to things like parameter norm growth and noise in gradients, tweaking hyperparameters (like using scheduled weight decay) to keep the training process stable. The takeaway for researchers: making novel architectures work often requires this kind of diligent debugging and fine-tuning of the training process.</p><p><strong>Efficient Scaling with &#956;P (Mu Parametrization):</strong> Training six different model sizes could have been six times the work, but Falcon-H1 applied an advanced technique called <strong>Maximal Update Parametrization (&#956;P)</strong> to streamline this. &#956;P provides a theory-backed way to set hyperparameters (like learning rate) for one model size and then scale them to larger models reliably. Using &#956;P, the team found optimal settings on a base model and transferred them to all other sizes, allowing them to train <strong>multiple models in parallel without lots of per-model tuning</strong>. They even went a step further by customizing the &#956;P scaling: instead of assuming the base model&#8217;s hyperparameters were perfectly tuned, they <strong>optimized 35 separate &#956;P scaling factors</strong> for different parts of the network. This fine-grained approach squeezed extra performance out of each model. For the layperson, the result is that <strong>Falcon-H1 models were trained efficiently and in concert</strong>, making this massive undertaking a bit more manageable.</p><p>All these choices &#8211; from architectural design to training strategy &#8211; contributed to Falcon-H1&#8217;s success. The researchers effectively <strong>broke some rules</strong> and <strong>set new ones</strong>. It&#8217;s a reminder that in the race for bigger and better AI, raw scale isn&#8217;t the only path; sometimes <strong>clever engineering and fresh ideas</strong> can unlock performance that brute force alone would miss.</p><h2>Results: When Smaller Models Beat Giants</h2><p>So, did the hybrid approach pay off? The proof is in the pudding (or rather, in a slew of benchmarks). The Falcon-H1 models have demonstrated <strong>state-of-the-art or near-SOTA performance</strong> across a variety of tasks, often rivaling models far larger:</p><p>The flagship <strong>Falcon-H1-34B</strong> model is reported to <strong>match or outperform systems in the 70B parameter range</strong>. In evaluations, it stood toe-to-toe with models like <strong>Qwen-72B (Alibaba&#8217;s model) and a prototype LLaMA-70B</strong> on many benchmarks. This is astonishing when you consider it has half (or even half to one-third) the parameters &#8211; a testament to <em>parameter efficiency</em>. On standard NLP benchmarks (like MMLU for knowledge, HellaSwag for common-sense reasoning, GSM8k for math, etc.), Falcon-34B&#8217;s numbers are right up there with the best, and sometimes <em>winning</em>&#12304;25&#8224;image&#12305;. It&#8217;s not uniformly dominant, but the fact that it&#8217;s in the same league as models twice its size is a huge win for the approach.</p><p>The <strong>Falcon-H1-1.5B-Deep</strong> model (just 1.5B params) deserves a shout-out. It <strong>clearly outperforms other ~2B models</strong> (like Qwen3-1.7B) and even <strong>holds its own against many 7B models</strong>. Essentially, a well-designed 1.5B hybrid model from 2025 can do what a vanilla 7B model from 2024 could do. That&#8217;s a big deal for those who need smaller models &#8211; it means you might not need a massive cluster to get strong AI performance for many tasks. Falcon-1.5B-Deep even beat some 7B versions of Falcon&#8217;s <em>own</em> previous generation (Falcon-40B from 2023 had a 7B little sibling) and competitive models, which validates the &#8220;deeper not wider&#8221; strategy in practice.</p><p>On <strong>multilingual benchmarks</strong>, Falcon-34B shows very strong performance across languages like Arabic, Spanish, Hindi, etc., often on par with the best models evaluated for those languages. This confirms that the model&#8217;s multilingual training wasn&#8217;t just token window dressing; it genuinely learned multiple languages well. For businesses in non-English markets or operating globally, this is a welcome capability &#8211; an AI that doesn&#8217;t treat English as the default brain.</p><p>In terms of <strong>reasoning and specialized tasks</strong>, Falcon-H1 models excel in categories like math and coding. The inclusion of math and code data gave them a leg up on benchmarks like GSM8k (math word problems) and HumanEval (coding tasks). For instance, Falcon-34B performs strongly on GSM8k and is competitive on coding challenges, approaching the scores of much larger models. The team even noted that they did <em>no special fine-tuning for reasoning</em>, yet the instruction-tuned models show robust reasoning abilities out-of-the-box. This is a promising sign that massive fine-tune rounds (like those done for GPT-4&#8217;s &#8220;Chain-of-Thought&#8221; reasoning) might be sidestepped with a good pretraining mix.</p><p>On <strong>long-context tasks</strong>, Falcon-H1 really shines. They tested the 34B model on retrieval and long-document QA tasks where it had to read very lengthy texts (tens of thousands of tokens) and answer questions. Its performance was strong and in some cases it could handle things that a typical model (with a 4K or 8K limit) simply couldn&#8217;t attempt. This hints at new applications: think AI assistants that can digest entire books or analyze years of financial reports in one go. One can imagine loading up a 200-page contract and asking the model detailed questions &#8211; Falcon-H1&#8217;s long attention span is built for that.</p><p>To sum up the results: <strong>Falcon-H1 proves that smart design can yield outsized results</strong>. A smaller open-source model, built on novel ideas, is reaching parity with the closed giants of the field. It&#8217;s not every day that David goes up against Goliath in AI and comes out looking this good. And importantly, <em>anyone</em> can use David &#8211; the models are downloadable, finetunable, and deployable by anyone with the hardware and savvy to do so.</p><h2>Why This Matters: A Glimpse of AI&#8217;s Future (Hybrid and Multimodal?)</h2><p>Beyond the cool factor of Falcon-H1&#8217;s achievements, there&#8217;s a broader significance here. This research is a signpost for several big trends in AI:</p><p><strong>1. Rethinking &#8220;Bigger is Better&#8221;:</strong> For the last few years, the story of AI has been &#8220;scale, scale, scale.&#8221; More parameters, more data, more compute &#8211; and you&#8217;ll get better results. That&#8217;s largely held true (GPT-4 wouldn&#8217;t be GPT-4 without an astronomical scale-up). But Falcon-H1 shows another path: <em>architectural innovation</em> and <em>training strategy</em> can give you leaps in performance without merely doubling size. It&#8217;s a reminder that we may be entering an era of <strong>diminishing returns on brute-force scaling</strong>, and the next breakthroughs could come from making models <strong>smarter, not just larger</strong>. For AI researchers, this is an exciting validation of pursuing new model forms (like combining attention with SSMs, or other hybrids) &#8211; there&#8217;s room to beat the scaling curve by being clever.</p><p><strong>2. The Rise of Modular &amp; Multimodal Foundation Models:</strong> Falcon-H1&#8217;s hybrid approach hints at a future where AI systems aren&#8217;t monolithic slabs of one architecture, but a combination of specialized components working together. Today it&#8217;s attention + SSM. Tomorrow it could be attention + SSM + convolution, or text + vision + memory modules, etc. In other words, <strong>multimodal foundation models</strong> might internally look like a <em>collection of expert modules</em> &#8211; an image understanding module, a text reasoning module, maybe a database retrieval module &#8211; all integrated in a single system. We&#8217;re already seeing early signs: OpenAI&#8217;s GPT-4 is <em>multimodal</em> (it can see images as well as read text), and it likely achieves this by combining a vision encoder with a language model. Google&#8217;s research has been exploring &#8220;<strong>universal models</strong>&#8221; that handle text, images, and more in one network. The success of Falcon-H1&#8217;s dual-technique model is a vote of confidence for this direction. It shows that <strong>heterogeneous architectures can work at scale</strong>. For tech strategists, this implies that the AI solutions of the near future might be more flexible and tailored &#8211; not every company will need a 100B-parameter pure Transformer if a 10B hybrid with the right components does the job better.</p><p><strong>3. Democratization of AI Capability:</strong> The open-source nature of projects like Falcon-H1 means cutting-edge AI is not confined to a few big tech companies. We&#8217;re witnessing a fast-follower effect where open models replicate or even innovate beyond the proprietary ones, often within months. This shrinks the advantage gap. For product builders and startups, it&#8217;s great news: you can pick up a state-of-the-art foundation model (with support for your language, your long documents, etc.) <strong>for free</strong> and fine-tune it to your needs. Running a 34B model that competes with a 70B might be feasible on off-the-shelf hardware, especially with quantization and optimizations. That lowers the barrier to entry for AI-powered products. It also means more room for customization &#8211; you&#8217;re not stuck with a one-size-fits-all API from a large provider; you can have a model in-house that you understand and control. For enterprise strategists, the question shifts from &#8220;<em>Where do we buy our AI capabilities?</em>&#8221; to &#8220;<em>How can we leverage these open foundation models to build unique value?</em>&#8221;. The playing field is leveling, and the competitive edge will go to those who can best adapt and integrate these rapidly improving open models.</p><p><strong>4. New Applications Unlocked:</strong> The technical improvements aren&#8217;t just academic; they translate to new use cases. <strong>256K context</strong> means AI that can <strong>digest entire knowledge bases or code repositories</strong> at once &#8211; imagine assistants that truly &#8220;know&#8221; your company&#8217;s documentation or an AI that can refactor a whole codebase in one shot. <strong>Multilingual fluency</strong> means one model can serve users across markets, or analyze text in multiple languages without separate systems. <strong>Better math and reasoning</strong> means more reliable analytical assistants (maybe it won&#8217;t replace your junior analyst, but it can double-check their work). These capabilities get product people thinking: what can we build now that was impractical before? Perhaps AI tutors that read and critique <em>textbooks</em>, legal AI that cross-references entire law libraries, or business intelligence bots that comb through years of reports and data dumps to answer high-level questions. We&#8217;re inching closer to AI that isn&#8217;t just a clever chatbot, but a genuine research assistant that can handle breadth and depth of information.</p><p>To bring it down to specifics, let&#8217;s consider a few perspectives:</p><p><strong>For Product Builders:</strong> Falcon-H1 and models like it mean you can deliver advanced AI features without needing a supercomputer or a mega-cloud budget. Want a customer support AI that handles <strong>long, complex threads</strong> of emails? A 7B or 34B Falcon can do that with its long context window. Need a coding helper that fits in your IDE? A 1.5B-Deep model might give surprisingly good code suggestions without calling an API. And because it&#8217;s open source, you can fine-tune it on your proprietary data (say, your company&#8217;s internal docs or codebase) with no external data leakage. The efficiency gain &#8211; doing more with smaller models &#8211; also hints at <strong>on-device AI</strong> for specialized apps. We could soon see smartphones or AR glasses running powerful language models locally, tailored to the user&#8217;s own data, thanks to these optimizations.</p><p><strong>For AI Researchers:</strong> Falcon-H1 is a case study in breaking the mold. It encourages researchers to explore <strong>hybrid architectures</strong> and not assume Transformers are the end-all be-all. The success with SSMs may reinvigorate research into other sequence models or into better ways of integrating modules (maybe techniques from control theory or neuroscience-inspired models could be the next plug-in component). Also, the team&#8217;s willingness to challenge training orthodoxies (like reverse curriculum and data reuse) is a reminder that <strong>empirical results can defy expectations</strong> &#8211; we should keep testing those assumptions, especially as we venture into regimes (like 10+ trillion token training) where our old intuitions might not hold. In short, the field may become less uniform; expect a Cambrian explosion of model variants, each with different mixtures of components aimed at different niches (language, vision, speech, etc.), all coexisting and advancing the state of the art in tandem.</p><p><strong>For Tech Strategists and Business Leaders:</strong> The broader trend indicated by this work is that <strong>AI is becoming more accessible and more adaptable</strong>. The power that once required a fortune in compute and a crack research team is rapidly trickling down. This compresses the timeline of AI adoption across industries &#8211; your competitors might deploy GPT-4-level AI in their workflows via an open model well before you budget for a big vendor contract. It also suggests that proprietary advantage in AI can be short-lived; an algorithmic breakthrough today might be open-sourced tomorrow. However, it also opens opportunities: organizations can <strong>craft their own foundation models</strong> tuned to their domain (be it finance, biomedicine, law) by starting from something like Falcon-H1 and extending it. Strategic differentiation may come from <em>how</em> you use and fine-tune the foundation models, rather than who has the biggest model. Additionally, keep an eye on the <strong>multimodal</strong> aspect &#8211; as models begin to handle text, images, and more in one system, companies that leverage that (e.g., an AI that can read documents and analyze associated charts/graphs all at once) will have an edge in insight extraction and automation.</p><div><hr></div><p><strong>Takeaway:</strong> The Falcon-H1 project shows that the future of AI won&#8217;t just be about piling on more parameters &#8211; it will be about <em>combining ideas and domains</em> to create smarter, more efficient systems. A hybrid model with &#8220;two brains&#8221; (attention + SSM) can outperform much larger single-minded models, hinting that the next leaps in AI might come from this spirit of integration. For those of us building and using these technologies, it&#8217;s a thrilling development. We should be prepared for AI models that are <strong>more like toolkits</strong> than one-trick ponies &#8211; able to see, listen, remember vast contexts, and reason, all in one. The foundations of AI are evolving, and with efforts like Falcon-H1, we&#8217;re getting a preview of a more <strong>surprising, accessible, and multimodal</strong> AI era to come.</p><div><hr></div><h3>&#10084;&#65039; If you enjoyed this article, give it a like and share it with your peers.</h3>]]></content:encoded></item><item><title><![CDATA[Can AI Really Understand How We Think? ]]></title><description><![CDATA[Controversial "Centaur" Model Sparks Fierce Debate]]></description><link>https://www.llmwatch.com/p/can-ai-really-understand-how-we-think</link><guid isPermaLink="false">https://www.llmwatch.com/p/can-ai-really-understand-how-we-think</guid><dc:creator><![CDATA[Pascal Biese]]></dc:creator><pubDate>Sat, 05 Jul 2025 14:31:11 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/e4e67fd4-352a-4711-9e76-5bcba07a5bd2_1280x720.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>You've probably noticed how AI seems to be getting eerily good at predicting human behavior. From recommendation algorithms that know what you'll want to watch next to chatbots that can mimic your writing style, these systems are becoming uncannily accurate at anticipating our choices. But the million-dollar question remains: does predicting behavior mean understanding cognition?</p><p>The debate around this question just exploded into a major scientific controversy with the publication of the Centaur model in <a href="https://www.nature.com/articles/s41586-025-09215-4">Nature</a>. Researchers at the Max Planck Institute fine-tuned Meta's Llama 3.1 language model on a massive dataset of human behavioral experiments, creating what they claim is a "unified model of human cognition." It didn't take long for the backlash from cognitive scientists to be manifested.</p><p>What we&#8217;re going to cover in this article:</p><ol><li><p>What exactly the Centaur model is and why it's causing such a stir</p></li><li><p>The impressive (and potentially concerning) capabilities it demonstrates</p></li><li><p>Why leading cognitive scientists are calling it "absurd"</p></li><li><p> What this means for the future of understanding the human mind</p></li><li><p> The genuinely valuable contributions hidden beneath the controversy</p></li></ol><p>Let's unpack what could be the most divisive scientific paper that has been published this year.</p>
      <p>
          <a href="https://www.llmwatch.com/p/can-ai-really-understand-how-we-think">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[Scenario Testing: A New Paradigm for Making AI Agents More Reliable]]></title><description><![CDATA[Ship autonomous agents with confidence, not crossed fingers]]></description><link>https://www.llmwatch.com/p/scenario-testing-a-new-paradigm-for</link><guid isPermaLink="false">https://www.llmwatch.com/p/scenario-testing-a-new-paradigm-for</guid><dc:creator><![CDATA[Pascal Biese]]></dc:creator><pubDate>Mon, 30 Jun 2025 15:30:25 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/9342c631-f763-4ab6-ac3c-3b838d958d21_1280x720.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>The rise of AI agents marks a <a href="https://www.youtube.com/watch?v=LCEmiRjPEtQ">fundamental shift</a> in software development. Unlike traditional code that follows predictable paths, AI agents show emergent behaviors, make autonomous decisions, and engage in complex multi-turn conversations. Yet most teams still test these sophisticated systems with manual, ad-hoc approaches - typing test messages, hoping they've covered edge cases, and crossing their fingers before deployment. This testing gap is becoming critical: AI-related incidents increased <strong>56.4% in 2024</strong>, and Gartner <a href="https://www.gartner.com/en/newsroom/press-releases/2025-06-25-gartner-predicts-over-40-percent-of-agentic-ai-projects-will-be-canceled-by-end-of-2027">predicts</a> <strong>40% of agentic AI projects will be cancelled by 2027</strong> due to insufficient testing and unclear ROI.</p><p>There must be a better way. Just as we use compilers to check code and linters to enforce standards, we need AI-native testing approaches for AI agents. One possible solution to this problem is scenario testing - a paradigm where AI agents systematically test other AI agents through realistic conversations and scenarios. LangWatch, an <a href="https://github.com/langwatch/langwatch">open-source</a> LLMOps platform, has pioneered this approach with a framework that transforms how teams validate AI agent behavior.</p><h2>The fundamental challenge of testing non-deterministic systems</h2><p>Traditional software testing relies on deterministic behavior: given input X, expect output Y. This approach breaks down completely with AI agents. Consider a customer service agent handling a billing inquiry - the same question might generate different phrasings, follow different conversation paths, or use different tools to reach the same outcome. The challenge isn't just variability; it's that <strong>the variability itself is a feature</strong>, allowing agents to handle novel situations and adapt to user needs.</p><p>This creates three critical testing challenges that manual approaches can't solve:</p><p><strong>Coverage blindness</strong>: Human testers inevitably test happy paths and obvious edge cases, missing the long tail of unexpected interactions. When an AI agent encounters "I need to speak to your manager, but first, can you explain why my bill shows charges from last Tuesday when I was in the hospital?"- has anyone tested that specific scenario?</p><p><strong>Regression invisibility</strong>: A seemingly minor prompt change to improve one interaction can break five others. Without systematic testing, these regressions only surface in production, often through frustrated user reports or, worse, silent failures that erode trust.</p><p><strong>Scale impossibility</strong>: As agents become more sophisticated - handling complex workflows, using multiple tools, maintaining context across sessions - the combinatorial explosion of possible interactions makes manual testing impractical. Testing a simple 5-turn conversation with 10 possible user responses at each turn creates 100,000 potential paths.</p><h2>Scenario testing: Simulation at scale</h2><p>A key insight of the <a href="https://huggingface.co/learn/cookbook/en/llm_judge">LLM-as-a-judge</a> method was recognizing that <strong>the best way to test LLMs at scale is with another LLM. </strong>While &#8220;judging&#8221; can be a very intransparent, complex process in itself, it&#8217;s much easier to evaluate the success of a well-defined workflow than, let&#8217;s say, the quality of the output (which is what the objective of LLM-as-a-judge methods usually is). So how are we <strong>testing AI agents with another AI agent</strong>? Instead of humans manually typing test messages, the current scenario testing approach uses a three-agent architecture:</p><ol><li><p><strong>Your agent</strong>: The AI system being tested</p></li><li><p><strong>User simulator agent</strong>: An AI that realistically simulates user behavior based on scenario descriptions</p></li><li><p><strong>Judge agent</strong>: An AI that evaluates whether the conversation met defined success criteria</p></li></ol><p>This approach transforms testing from manual execution to scenario definition. Let&#8217;s take a look at an example from the LangWatch repository - rather than scripting exact conversations, teams describe situations and desired outcomes in natural language:</p><pre><code><code>result = await scenario.run(
    name="billing_inquiry_edge_case",
    description="""
    A frustrated customer received a bill with unexplained charges. 
    They're skeptical of automated systems and may challenge the agent's responses.
    They have a valid concern about a duplicate charge from last month.
    """,
    agents=[
        CustomerServiceAgent(),
        scenario.UserSimulatorAgent(),
        scenario.JudgeAgent(
            criteria=[
                "Agent acknowledges customer frustration empathetically",
                "Agent identifies the duplicate charge without defensive language",
                "Agent offers concrete resolution within policy guidelines",
                "Conversation resolves within 5 exchanges"
            ]
        )
    ]
)
</code></code></pre><p>The user simulator agent interprets this description and dynamically generates realistic user messages, adapting based on how the conversation evolves. The judge agent evaluates the entire conversation against the success criteria, providing both pass/fail results and detailed analysis.</p><h2>The multiplicative power of automated scenario testing</h2><p>This simulation-based approach unlocks capabilities impossible with manual testing:</p><p><strong>Systematic edge case exploration</strong>: Teams can define hundreds of scenarios covering edge cases, adversarial inputs, and complex multi-step workflows. Each scenario runs automatically, exploring different conversation paths while maintaining consistent evaluation criteria.</p><p><strong>Continuous regression detection</strong>: Integrate scenario tests into CI/CD pipelines to run on every code change. That prompt tweak that seemed harmless? Scenario tests catch the three customer journeys it inadvertently broke before they reach production.</p><p><strong>Domain expert collaboration</strong>: Product managers and subject matter experts can define scenarios and success criteria without writing code. A healthcare product manager can specify: "When a patient asks about drug interactions, the agent must always recommend consulting their physician" without understanding the underlying implementation.</p><p><strong>Parallel testing at scale</strong>: Unlike human testers who process conversations sequentially, automated scenarios run in parallel. Test suites with hundreds of scenarios complete in minutes, not days.</p><p>Consider a real example: a financial services company deployed an AI agent to handle loan inquiries. Manual testing covered basic flows - checking rates, application status, payment questions. But scenario testing revealed critical gaps:</p><ul><li><p>The agent provided different interest rates when asked in slightly different ways</p></li><li><p>Multi-part questions about refinancing options confused the agent's context management</p></li><li><p>Edge cases around regulatory disclosures were inconsistently handled</p></li></ul><p>Each issue represented not just a bug, but a compliance risk. Scenario testing caught them before they became regulatory violations.</p><div><hr></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.llmwatch.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">LLM Watch is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><div><hr></div><h2>Beyond pass/fail: The strategic value of scenario testing</h2><p>The immediate benefit of catching bugs is obvious, but scenario testing delivers deeper strategic value:</p><p><strong>Behavioral contracts</strong>: Scenarios codify expected agent behavior, creating living documentation of system capabilities. When stakeholders ask "Can our agent handle X?", the scenario suite provides definitive answers.</p><p><strong>Confident iteration</strong>: Teams can improve agents rapidly, knowing that scenario tests catch regressions. This accelerates the development cycle from cautious monthly releases to confident daily deployments.</p><p><strong>Quality metrics</strong>: Track scenario pass rates over time to measure agent improvement. Use failure analysis to identify systematic issues - if multiple scenarios fail around date handling, you've identified a capability gap.</p><p><strong>Cost optimization</strong>: Every production issue caught in testing saves exponentially more than the compute cost of running scenarios. Organizations report <strong>80% reduction in production incidents</strong> after implementing comprehensive scenario testing.</p><h2>Implementation patterns for successful scenario testing</h2><p>Based on analysis of successful implementations, several patterns emerge:</p><h3>Start with critical paths, not comprehensive coverage</h3><p>Resist the temptation to test everything immediately. Begin with your highest-risk, highest-value user journeys:</p><pre><code><code># Start here: Core business-critical scenarios
- Customer purchase flow
- Account access recovery  
- Complaint escalation handling

# Not here: Every possible greeting variation
</code></code></pre><h3>Design scenarios for insight, not just validation</h3><p>Good scenarios reveal how agents behave under stress:</p><pre><code><code># Less useful: Happy path validation
description = "User asks for store hours"

# More useful: Stress testing
description = """
User is trying to make a return after hours on the last day 
of the return window. They're frustrated and mention considering 
switching to a competitor. They have a receipt but it's partially damaged.
"""
</code></code></pre><h3>Layer your testing strategy</h3><p>Scenario testing complements, not replaces, other testing approaches:</p><ol><li><p><strong>Unit tests</strong>: Validate individual functions and tools</p></li><li><p><strong>Component evals</strong>: Test specific capabilities (e.g., information extraction)</p></li><li><p><strong>Scenario tests</strong>: Validate end-to-end user journeys</p></li><li><p><strong>Production monitoring</strong>: Track real-world performance</p></li></ol><h3>Embrace non-determinism thoughtfully</h3><p>Rather than fighting AI variability, design scenarios that accommodate it:</p><pre><code><code>criteria = [
    # Too rigid: "Agent must say 'Thank you for contacting support'"
    
    # Better: "Agent acknowledges the customer's contact professionally"
    # Better: "Agent provides refund amount between $47-53 based on calculation"
    # Better: "Resolution offered aligns with company policy document"
]
</code></code></pre><h3>Enable cross-functional collaboration</h3><p>The most successful implementations involve diverse perspectives:</p><ul><li><p><strong>Engineers</strong> define technical constraints and integration requirements</p></li><li><p><strong>Product managers</strong> specify user journeys and success metrics</p></li><li><p><strong>Domain experts</strong> contribute realistic scenarios and evaluation criteria</p></li><li><p><strong>QA teams</strong> organize test suites and analyze failure patterns</p></li></ul><h2>Common pitfalls and how to avoid them</h2><h3>The over-automation trap</h3><p>Not every test needs to be a scenario. Simple deterministic checks remain valuable:</p><pre><code><code># Overkill: Scenario test for API availability
# Better: Simple unit test

# Appropriate: Scenario test for multi-step booking flow
# Overkill: Unit test for conversation dynamics
</code></code></pre><h3>The perfect scenario fallacy</h3><p>Scenarios don't need to cover every possibility - they need to cover meaningful possibilities:</p><pre><code><code># Too broad: "User asks about products"
# Too narrow: "User asks about blue widgets on Tuesday afternoon"
# Just right: "Price-sensitive customer comparing products for a home renovation"
</code></code></pre><h3>The isolation error</h3><p>Testing agents in isolation misses integration issues. Include scenarios that exercise:</p><ul><li><p>Tool usage and API calls</p></li><li><p>Knowledge base retrieval</p></li><li><p>Multi-agent handoffs</p></li><li><p>Session persistence</p></li></ul><h2>The path forward: Making scenario testing operational</h2><p>For teams ready to implement scenario testing, the most crucial first step is establishing a solid foundation with the right tooling and initial test coverage. Start by setting up a framework like <a href="https://github.com/langwatch/langwatch">LangWatch</a> or a similar platform that can handle the complexity of scenario-based testing. Rather than trying to cover everything at once, focus your initial efforts on creating just five to ten scenarios that cover your most critical user paths. These should be the workflows that, if broken, would have the most significant impact on your users or business.</p><p>During these first couple of weeks, integration with your existing test suite is essential. This isn't about replacing what you have but augmenting it with scenario-based coverage. Take time to establish clear team conventions around how scenarios should be written, named, and organized. This early investment in standards will pay dividends as your test suite grows.</p><p>Once you have your foundation in place, the natural progression is to expand your coverage to include edge cases and error paths. This is where scenario testing really shines compared to traditional unit tests. You'll want to implement caching mechanisms to keep your tests running efficiently as the suite grows. Getting your scenarios integrated into your CI/CD pipeline at this stage ensures that they become part of your regular development workflow rather than an afterthought.</p><p>The key to long-term success with scenario testing is treating it as a living system rather than a one-time implementation. Regular reviews of your scenarios ensure they stay relevant as your application evolves. Analyzing failure patterns helps you identify areas where your application might be fragile, while performance benchmarking ensures your tests remain fast enough to provide quick feedback. Perhaps most importantly, creating mechanisms for knowledge sharing across teams helps spread both the benefits and the learnings from your scenario testing efforts throughout the organization.</p><p>Remember that scenario testing is most effective when it becomes part of your team's culture rather than just another checkbox in your process. Start small, focus on value, and let the practice grow organically as your team sees the benefits firsthand.</p><div><hr></div><h3><strong>&#10084;&#65039; If you enjoyed this article, give it a like and share it with your peers.</strong></h3>]]></content:encoded></item><item><title><![CDATA[Bursting the "AI Is Just Memorization"-Bubble]]></title><description><![CDATA[A Deep Dive into Measuring LLM Capacity]]></description><link>https://www.llmwatch.com/p/bursting-the-ai-is-just-memorization</link><guid isPermaLink="false">https://www.llmwatch.com/p/bursting-the-ai-is-just-memorization</guid><dc:creator><![CDATA[Pascal Biese]]></dc:creator><pubDate>Wed, 04 Jun 2025 15:02:24 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/820483b6-bcbe-4089-b308-3e2086f06863_1536x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Recent advancements in large language models (LLMs) - and more generally, Generative AI - have sparked intense debate about data memorization. Can these models reproduce their training data verbatim? How much information do they actually store? And perhaps most importantly, when does beneficial learning end and problematic memorization begin?</p><p>A research coalition consisting of researchers from Meta's FAIR, Google DeepMind, Cornell, and NVIDIA set out to shed some light on these questions - and with success, it seems.  Their new paper titled "<a href="https://www.arxiv.org/abs/2505.24832">How much do language models memorize?</a>" provides a rigorous mathematical framework for measuring memorization and delivers surprising insights about the fundamental capacity limits of transformer models.</p><p>What we'll cover in this article:</p><ul><li><p>Why existing definitions of memorization fall short</p></li><li><p>A new compression-based framework for measuring memorization</p></li><li><p>The surprising discovery that GPT-style models store ~3.6 bits per parameter</p></li><li><p>How memorization relates to the double descent phenomenon</p></li><li><p>Practical scaling laws for membership inference attacks</p></li><li><p>What this means for the future of LLM development</p></li></ul><p>Are you ready? Let's dive in.</p><h2>1. The Memorization Problem: Why Current Definitions Don't Work</h2><p>Before we can measure how much models memorize, we need to define what memorization actually means. This turns out to be surprisingly tricky.</p><h3>The Extraction Fallacy</h3><p>Most existing work defines memorization through extraction: if you can prompt a model to generate a specific training sequence, it must have memorized it. But the authors point out a critical flaw in this reasoning. Modern LLMs can be coerced to output almost any string with the right prompt. As they note, "the fact that a model outputs something is not necessarily a sign of memorization."</p><p>Consider this example: if you prompt a model with "What is 2^100?" and it correctly responds with "1,267,650,600,228,229,401,496,703,205,376", has it memorized this specific fact, or has it learned to perform exponentiation? The extraction-based definition can't distinguish between these fundamentally different scenarios.</p><h3>The Stability Problem</h3><p>Other definitions rely on differential privacy or influence functions, measuring how a model changes when a training example is added or removed. But these approaches have their own limitations:</p><ul><li><p>They depend heavily on the training algorithm</p></li><li><p>They measure worst-case behavior rather than typical memorization</p></li><li><p>They can't be applied to a single model in isolation</p></li></ul><p>The authors needed something different - a definition that could separate memorization from generalization, work at the sample level, and be independent of the training process.</p>
      <p>
          <a href="https://www.llmwatch.com/p/bursting-the-ai-is-just-memorization">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[AlphaEvolve: Google DeepMind's Latest Breakthrough Success]]></title><description><![CDATA[A Coding Agent for Scientific and Algorithmic Discovery]]></description><link>https://www.llmwatch.com/p/alphaevolve-google-deepminds-latest</link><guid isPermaLink="false">https://www.llmwatch.com/p/alphaevolve-google-deepminds-latest</guid><dc:creator><![CDATA[Pascal Biese]]></dc:creator><pubDate>Thu, 15 May 2025 14:15:53 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/0da29f15-5937-48d9-befb-8f6b5d36df72_1280x720.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>In order to push the boundaries of computational capabilities, researchers have long sought automated methods to discover novel and improved algorithms. The recently announced <a href="https://storage.googleapis.com/deepmind-media/DeepMind.com/Blog/alphaevolve-a-gemini-powered-coding-agent-for-designing-advanced-algorithms/AlphaEvolve.pdf">AlphaEvolve</a> from Google DeepMind brings us one step closer to achieving this dream. Their evolutionary coding agent combines the pattern recognition and code generation capabilities of Large Language Models (LLMs) with evolutionary computation to tackle some of the most challenging problems in computer science and mathematics.</p><p>What distinguishes AlphaEvolve from previous approaches is its ability to evolve entire codebases rather than just single functions, work across multiple programming languages, and leverage rich contextual information to guide the evolutionary process. These capabilities have enabled breakthroughs in longstanding mathematical problems and meaningful optimizations in critical computational infrastructure.</p><p>In this deep dive, we'll explore how AlphaEvolve works, what makes it different from previous approaches, and examine its impressive results across scientific discovery and practical engineering applications.</p><div><hr></div><h2>TL;DR - Technical Innovations</h2><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Bl4U!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2d247b6-4d32-4642-92d8-7069f1928f87_1133x440.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Bl4U!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2d247b6-4d32-4642-92d8-7069f1928f87_1133x440.png 424w, https://substackcdn.com/image/fetch/$s_!Bl4U!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2d247b6-4d32-4642-92d8-7069f1928f87_1133x440.png 848w, https://substackcdn.com/image/fetch/$s_!Bl4U!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2d247b6-4d32-4642-92d8-7069f1928f87_1133x440.png 1272w, https://substackcdn.com/image/fetch/$s_!Bl4U!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2d247b6-4d32-4642-92d8-7069f1928f87_1133x440.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Bl4U!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2d247b6-4d32-4642-92d8-7069f1928f87_1133x440.png" width="1133" height="440" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b2d247b6-4d32-4642-92d8-7069f1928f87_1133x440.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:440,&quot;width&quot;:1133,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:99416,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.llmwatch.com/i/163618315?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2d247b6-4d32-4642-92d8-7069f1928f87_1133x440.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!Bl4U!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2d247b6-4d32-4642-92d8-7069f1928f87_1133x440.png 424w, https://substackcdn.com/image/fetch/$s_!Bl4U!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2d247b6-4d32-4642-92d8-7069f1928f87_1133x440.png 848w, https://substackcdn.com/image/fetch/$s_!Bl4U!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2d247b6-4d32-4642-92d8-7069f1928f87_1133x440.png 1272w, https://substackcdn.com/image/fetch/$s_!Bl4U!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb2d247b6-4d32-4642-92d8-7069f1928f87_1133x440.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>AlphaEvolve represents a substantial advancement over previous LLM-guided evolution systems like FunSearch. Key improvements include:</p><ol><li><p><strong>Scale and scope</strong>: While FunSearch only evolved single Python functions of 10-20 lines, AlphaEvolve can evolve entire code files with hundreds of lines in any programming language.</p></li><li><p><strong>Evaluation capabilities</strong>: AlphaEvolve can handle evaluations running for hours on accelerators, compared to FunSearch's limitation of &#8804;20 minutes on a single CPU.</p></li><li><p><strong>Sample efficiency</strong>: AlphaEvolve requires only thousands of LLM samples rather than millions.</p></li><li><p><strong>Model utilization</strong>: AlphaEvolve benefits significantly from state-of-the-art LLMs, whereas FunSearch showed minimal benefit from larger models.</p></li><li><p><strong>Context richness</strong>: AlphaEvolve uses rich context and feedback in prompts, beyond just previous solutions.</p></li><li><p><strong>Multi-objective optimization</strong>: AlphaEvolve can simultaneously optimize multiple metrics, not just a single objective.</p></li></ol><p>Ablation studies confirmed the importance of each component:</p><ul><li><p>The evolutionary approach (versus repeatedly feeding the same initial program to an LLM)</p></li><li><p>Rich context in prompts</p></li><li><p>Meta-prompt evolution</p></li><li><p>Full-file evolution capability</p></li><li><p>The use of powerful language models</p></li></ul><p>Each of these components contributed significantly to AlphaEvolve's performance across different tasks. </p><p>So let&#8217;s take a closer look at how the systems works, what kind of breakthrough solutions it produced exactly, and how Google was able to put them to use.</p>
      <p>
          <a href="https://www.llmwatch.com/p/alphaevolve-google-deepminds-latest">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[Llama-Nemotron: NVIDIA's Foundation Model for Agentic AI]]></title><description><![CDATA[A New Generation of Efficient Reasoning Models]]></description><link>https://www.llmwatch.com/p/llama-nemotron-nvidias-foundation</link><guid isPermaLink="false">https://www.llmwatch.com/p/llama-nemotron-nvidias-foundation</guid><dc:creator><![CDATA[Pascal Biese]]></dc:creator><pubDate>Thu, 08 May 2025 14:04:27 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/786eb472-ad80-4434-b4c4-814cd77e45e8_1536x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>In recent months, the emergence of &#8220;reasoning&#8221;-optimized Large Language Models (LLMs)  models capable of emitting multi-step chains of thought, self-verification, and backtracking - has reshaped what we expect from AI assistants. However, powering these capabilities at scale still poses a challenge: long, compute-intensive inference runs can become prohibitively expensive, and a one-size-fits-all reasoning strategy is not always ideal. </p><p>NVIDIA&#8217;s newly released <a href="https://arxiv.org/abs/2505.00949">Llama-Nemotron</a> (LN) family addresses these issues, delivering models that (1) support a user-controllable reasoning toggle, (2) pack state-of-the-art scientific and mathematical reasoning into footprints that fit on commodity hardware, and (3) offer open licenses for enterprise and research use.</p><p>In this deep dive, we will explore the architecture, training methodology, and innovations that make Llama-Nemotron stand out in an increasingly crowded landscape of LLMs.</p><h2><strong>Key Contributions in 30 Seconds</strong></h2><p>The Llama-Nemotron family introduces several notable architecture decisions:</p><ol><li><p><strong>Heterogeneous architecture</strong> optimized for inference efficiency through neural architecture search</p></li><li><p><strong>Dynamic reasoning toggle</strong> allowing users to switch between standard chat and reasoning modes</p></li><li><p><strong>FFN Fusion</strong> technique to reduce sequential depth and improve inference latency</p></li><li><p><strong>Large-scale reinforcement learning</strong> pushing reasoning capabilities beyond teacher models</p></li><li><p><strong>FP8 inference generation</strong> for significantly improved throughput</p></li></ol><p>The models come in three sizes - Nano (8B), Super (49B), and Ultra (253B) - each optimized for specific deployment scenarios while maintaining strong reasoning capabilities.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!w_RB!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff3fff46e-bc88-4831-a4e3-d44329487b6c_1175x586.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!w_RB!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff3fff46e-bc88-4831-a4e3-d44329487b6c_1175x586.png 424w, https://substackcdn.com/image/fetch/$s_!w_RB!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff3fff46e-bc88-4831-a4e3-d44329487b6c_1175x586.png 848w, https://substackcdn.com/image/fetch/$s_!w_RB!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff3fff46e-bc88-4831-a4e3-d44329487b6c_1175x586.png 1272w, https://substackcdn.com/image/fetch/$s_!w_RB!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff3fff46e-bc88-4831-a4e3-d44329487b6c_1175x586.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!w_RB!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff3fff46e-bc88-4831-a4e3-d44329487b6c_1175x586.png" width="1175" height="586" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f3fff46e-bc88-4831-a4e3-d44329487b6c_1175x586.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:586,&quot;width&quot;:1175,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:148422,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.llmwatch.com/i/163118463?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff3fff46e-bc88-4831-a4e3-d44329487b6c_1175x586.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!w_RB!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff3fff46e-bc88-4831-a4e3-d44329487b6c_1175x586.png 424w, https://substackcdn.com/image/fetch/$s_!w_RB!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff3fff46e-bc88-4831-a4e3-d44329487b6c_1175x586.png 848w, https://substackcdn.com/image/fetch/$s_!w_RB!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff3fff46e-bc88-4831-a4e3-d44329487b6c_1175x586.png 1272w, https://substackcdn.com/image/fetch/$s_!w_RB!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff3fff46e-bc88-4831-a4e3-d44329487b6c_1175x586.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">As of April 2025, LN-Ultra is the most &#8220;intelligent&#8221; open model according to Artificial Analysis. <a href="https://artificialanalysis.ai/">Source</a>.</figcaption></figure></div>
      <p>
          <a href="https://www.llmwatch.com/p/llama-nemotron-nvidias-foundation">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[ThinkPRM: More Than Just Chain-of-Thought (CoT 2.0)]]></title><description><![CDATA[AI models that verify reasoning steps with Advanced Chain-of-Thought]]></description><link>https://www.llmwatch.com/p/thinkprm-more-than-just-chain-of</link><guid isPermaLink="false">https://www.llmwatch.com/p/thinkprm-more-than-just-chain-of</guid><dc:creator><![CDATA[Pascal Biese]]></dc:creator><pubDate>Wed, 30 Apr 2025 12:59:40 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/7d3e0fc8-28a1-4727-ab29-24ef5481f0ea_1280x720.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>As Artificial Intelligence (AI) advances, large language models (LLMs) like ChatGPT and Claude have become increasingly capable of solving complex problems through step-by-step reasoning. But their solutions are only as valuable as they are accurate. The <a href="https://www.arxiv.org/abs/2504.16828">ThinkPRM</a> paper introduces a breakthrough approach to efficiently verify AI reasoning processes. This addresses a critical challenge in AI: how to reliably check if an AI's step-by-step reasoning is correct without requiring enormous amounts of human-labeled data.</p><h2><strong>Why Verification Is Hard</strong></h2><p>When an LLM solves a complex math problem or writes computer code, it produces a chain of reasoning steps. Ensuring these steps are correct is crucial for applications in education, scientific research, and critical decision-making. The traditional approach to this verification relies on process reward models (PRMs) &#8211; specialized AI systems that score each step in a solution.Until now, there have been two main verification approaches:</p><ol><li><p><strong>Discriminative PRMs</strong>: These models classify each reasoning step as correct or incorrect. They're effective but require massive datasets with step-by-step human annotations &#8211; often hundreds of thousands of labeled examples. Creating this data is time-consuming and expensive.</p></li><li><p><strong>LLM-as-a-Judge</strong>: This approach prompts an existing LLM to evaluate solutions without additional training. While convenient, these models often struggle with complex reasoning tasks and can produce unreliable results. They frequently suffer from problems like "overthinking" (generating excessively long verifications) or getting stuck in repetitive loops.</p></li></ol><p>Despite advances in both approaches, verification systems face persistent challenges. Discriminative PRMs depend on extensive labeled data that's costly to create, while LLM-as-a-Judge approaches often make errors in complex reasoning scenarios and struggle with consistency. These limitations have constrained progress in developing reliable verification systems that can handle sophisticated reasoning tasks efficiently.</p><h2><strong>ThinkPRM: A Potential Breakthrough</strong></h2><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!GXU4!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Feeb6d61f-8644-4e56-bef3-694b77f2b9d7_1103x605.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!GXU4!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Feeb6d61f-8644-4e56-bef3-694b77f2b9d7_1103x605.png 424w, https://substackcdn.com/image/fetch/$s_!GXU4!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Feeb6d61f-8644-4e56-bef3-694b77f2b9d7_1103x605.png 848w, https://substackcdn.com/image/fetch/$s_!GXU4!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Feeb6d61f-8644-4e56-bef3-694b77f2b9d7_1103x605.png 1272w, https://substackcdn.com/image/fetch/$s_!GXU4!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Feeb6d61f-8644-4e56-bef3-694b77f2b9d7_1103x605.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!GXU4!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Feeb6d61f-8644-4e56-bef3-694b77f2b9d7_1103x605.png" width="1103" height="605" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/eeb6d61f-8644-4e56-bef3-694b77f2b9d7_1103x605.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:605,&quot;width&quot;:1103,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:127292,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.llmwatch.com/i/162520549?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Feeb6d61f-8644-4e56-bef3-694b77f2b9d7_1103x605.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!GXU4!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Feeb6d61f-8644-4e56-bef3-694b77f2b9d7_1103x605.png 424w, https://substackcdn.com/image/fetch/$s_!GXU4!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Feeb6d61f-8644-4e56-bef3-694b77f2b9d7_1103x605.png 848w, https://substackcdn.com/image/fetch/$s_!GXU4!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Feeb6d61f-8644-4e56-bef3-694b77f2b9d7_1103x605.png 1272w, https://substackcdn.com/image/fetch/$s_!GXU4!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Feeb6d61f-8644-4e56-bef3-694b77f2b9d7_1103x605.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>ThinkPRM (Process Reward Models That Think) is a new approach that fundamentally reimagines verification as a generative, reasoning-based task rather than a simple classification problem.</p><p>It works by leveraging the inherent reasoning abilities of language models to verify reasoning. Instead of merely classifying steps as correct or incorrect, it "thinks through" each step, generating detailed verification chains-of-thought (CoT) that explain why a step is right or wrong.Here's the innovative process:</p><ol><li><p><strong>Foundation</strong>: The researchers start with open-source reasoning models like R1-Distill-Qwen.</p></li><li><p><strong>Synthetic Data Generation</strong>: Rather than requiring extensive human annotations, they prompt a larger language model (QwQ-32B-Preview) to generate verification chains for a sample of problem solutions.</p></li><li><p><strong>Quality Filtering</strong>: They only keep verification chains that match known step-level labels, ensuring high-quality training data.</p></li><li><p><strong>Lightweight Training</strong>: The model is fine-tuned on this small but high-quality dataset &#8211; just 1,000 carefully filtered examples (representing about 8,000 step-level annotations).</p></li></ol><p>The result is a model that can carefully analyze each step in a solution, explaining its reasoning process and providing a judgment about correctness. Here's an abbreviated example of ThinkPRM verifying a math problem solution:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!j_Ay!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff880e1bf-5e67-4e0d-a0d4-0f1564a71d68_1114x1124.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!j_Ay!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff880e1bf-5e67-4e0d-a0d4-0f1564a71d68_1114x1124.png 424w, https://substackcdn.com/image/fetch/$s_!j_Ay!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff880e1bf-5e67-4e0d-a0d4-0f1564a71d68_1114x1124.png 848w, https://substackcdn.com/image/fetch/$s_!j_Ay!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff880e1bf-5e67-4e0d-a0d4-0f1564a71d68_1114x1124.png 1272w, https://substackcdn.com/image/fetch/$s_!j_Ay!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff880e1bf-5e67-4e0d-a0d4-0f1564a71d68_1114x1124.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!j_Ay!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff880e1bf-5e67-4e0d-a0d4-0f1564a71d68_1114x1124.png" width="1114" height="1124" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f880e1bf-5e67-4e0d-a0d4-0f1564a71d68_1114x1124.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1124,&quot;width&quot;:1114,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:280636,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.llmwatch.com/i/162520549?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff880e1bf-5e67-4e0d-a0d4-0f1564a71d68_1114x1124.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!j_Ay!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff880e1bf-5e67-4e0d-a0d4-0f1564a71d68_1114x1124.png 424w, https://substackcdn.com/image/fetch/$s_!j_Ay!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff880e1bf-5e67-4e0d-a0d4-0f1564a71d68_1114x1124.png 848w, https://substackcdn.com/image/fetch/$s_!j_Ay!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff880e1bf-5e67-4e0d-a0d4-0f1564a71d68_1114x1124.png 1272w, https://substackcdn.com/image/fetch/$s_!j_Ay!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff880e1bf-5e67-4e0d-a0d4-0f1564a71d68_1114x1124.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>This "thinking" process provides transparency into the verification, making it easier to understand why a particular step is judged correct or incorrect.</p><div><hr></div><div class="digest-post-embed" data-attrs="{&quot;nodeId&quot;:&quot;21b04201-d310-4b3d-a901-91aa7f2453f9&quot;,&quot;caption&quot;:&quot;Multi-Agent Systems built using Large Language Models (LLMs) have emerged as a promising approach to complex problem-solving. By orchestrating multiple specialized agents working in concert, these systems aim to accomplish tasks that might be challenging for a single agent. However, despite growing enthusiasm and investment in these Multi-Agent LLM Systems (MAS), research reveals that their performance often falls short of expectations.&quot;,&quot;cta&quot;:&quot;Read full story&quot;,&quot;showBylines&quot;:true,&quot;size&quot;:&quot;md&quot;,&quot;isEditorNode&quot;:true,&quot;title&quot;:&quot;Multi-Agent Failure: What It Is and How to Prevent It&quot;,&quot;publishedBylines&quot;:[{&quot;id&quot;:48405812,&quot;name&quot;:&quot;Pascal Biese&quot;,&quot;bio&quot;:&quot;Human engineer reporting from the cutting edge of AI research.&quot;,&quot;photo_url&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/256e8620-1524-4496-a84d-7943a0edc098_512x512.png&quot;,&quot;is_guest&quot;:false,&quot;bestseller_tier&quot;:100}],&quot;post_date&quot;:&quot;2025-04-29T11:56:32.849Z&quot;,&quot;cover_image&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/05525bc3-3095-4854-9007-c276f2b2d6fc_1536x1024.png&quot;,&quot;cover_image_alt&quot;:null,&quot;canonical_url&quot;:&quot;https://www.llmwatch.com/p/multi-agent-failure-why-complex-ai&quot;,&quot;section_name&quot;:&quot;Deep Dives&quot;,&quot;video_upload_id&quot;:null,&quot;id&quot;:162316430,&quot;type&quot;:&quot;newsletter&quot;,&quot;reaction_count&quot;:11,&quot;comment_count&quot;:0,&quot;publication_id&quot;:null,&quot;publication_name&quot;:&quot;LLM Watch&quot;,&quot;publication_logo_url&quot;:&quot;https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5d95c476-43a7-4447-9081-9298a1fc325a_1280x1280.png&quot;,&quot;belowTheFold&quot;:true,&quot;youtube_url&quot;:null,&quot;show_links&quot;:null,&quot;feed_url&quot;:null}"></div><div><hr></div><h2><strong>Convincing Results</strong></h2><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!0u_S!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F55117820-0917-434b-87e6-a575a619cab7_1102x574.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!0u_S!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F55117820-0917-434b-87e6-a575a619cab7_1102x574.png 424w, https://substackcdn.com/image/fetch/$s_!0u_S!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F55117820-0917-434b-87e6-a575a619cab7_1102x574.png 848w, https://substackcdn.com/image/fetch/$s_!0u_S!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F55117820-0917-434b-87e6-a575a619cab7_1102x574.png 1272w, https://substackcdn.com/image/fetch/$s_!0u_S!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F55117820-0917-434b-87e6-a575a619cab7_1102x574.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!0u_S!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F55117820-0917-434b-87e6-a575a619cab7_1102x574.png" width="1102" height="574" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/55117820-0917-434b-87e6-a575a619cab7_1102x574.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:574,&quot;width&quot;:1102,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:144437,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.llmwatch.com/i/162520549?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F55117820-0917-434b-87e6-a575a619cab7_1102x574.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!0u_S!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F55117820-0917-434b-87e6-a575a619cab7_1102x574.png 424w, https://substackcdn.com/image/fetch/$s_!0u_S!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F55117820-0917-434b-87e6-a575a619cab7_1102x574.png 848w, https://substackcdn.com/image/fetch/$s_!0u_S!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F55117820-0917-434b-87e6-a575a619cab7_1102x574.png 1272w, https://substackcdn.com/image/fetch/$s_!0u_S!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F55117820-0917-434b-87e6-a575a619cab7_1102x574.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>ThinkPRM achieves exceptional performance while using drastically less training data:</p><h3><strong>1. Data Efficiency</strong></h3><p>Perhaps the most striking finding is that ThinkPRM performs better than discriminative PRMs trained on <strong>100 times more data</strong>. While traditional models required 700,000+ step-level annotations, the method achieves superior results with just 8,000 annotations.</p><h3><strong>2. Superior Performance</strong></h3><p>ThinkPRM outperforms both discriminative PRMs and LLM-as-a-judge approaches across multiple challenging benchmarks:</p><ul><li><p>On ProcessBench (a benchmark for identifying reasoning errors), ThinkPRM-14B achieves 86.5% F1 score compared to 73.7% for the LLM-as-a-judge approach using the same base model.</p></li><li><p>When used to guide search processes in solving MATH-500 problems, ThinkPRM-1.5B outperforms discriminative PRMs by approximately 5 percentage points.</p></li></ul><h3><strong>3. Generalization to New Domains</strong></h3><p>Despite being trained only on math problems, ThinkPRM demonstrates remarkable generalization to entirely different domains:</p><ul><li><p>On physics questions from GPQA-Diamond, ThinkPRM outperforms discriminative PRMs by 8 percentage points.</p></li><li><p>On code generation tasks from LiveCodeBench, it achieves a 4.5% advantage.</p></li></ul><p>This ability to generalize suggests that ThinkPRM is learning fundamental reasoning verification skills rather than domain-specific patterns.</p><h3><strong>4. Scalable Verification</strong></h3><p>A unique advantage of ThinkPRM is its ability to scale verification compute in two ways:</p><ol><li><p><strong>Parallel Scaling</strong>: Sampling multiple verification chains independently and aggregating their decisions improves accuracy by ~5 percentage points.</p></li><li><p><strong>Sequential Scaling</strong>: The model can "think longer" by extending its verification process, checking and revising its initial judgment. This capability allows ThinkPRM to continue improving as it's given more computation time.</p></li></ol><h2><strong>How This Changes AI Verification</strong></h2><p>ThinkPRM represents a fundamental shift in how we approach verification of complex reasoning:</p><h3><strong>From Classification to Reasoning</strong></h3><p>Traditional PRMs treat verification as a classification task &#8211; binary decisions about step correctness. ThinkPRM reframes verification as a reasoning task, where a model must think through and justify its evaluation. This approach is not just more data-efficient but also more aligned with how humans verify reasoning.</p><h3><strong>Transparency and Interpretability</strong></h3><p>Unlike black-box discriminative models, ThinkPRM's verification process is fully transparent. Users can read the model's verification chain to understand why it judged a step correct or incorrect. This transparency is crucial for applications where understanding the rationale behind verification decisions matters.</p><h3><strong>Low-Resource Adaptation</strong></h3><p>The remarkable data efficiency of ThinkPRM opens possibilities for creating specialized verifiers for niche domains where extensive labeled data is unavailable. This could democratize access to high-quality verification systems across diverse fields of expertise.</p><h2><strong>Challenges and Limitations</strong></h2><p>Despite its improvements, ThinkPRM still faces challenges:</p><ol><li><p><strong>Calibration</strong>: Like many LLMs, ThinkPRM can be overconfident, with scores clustering at extremes (near 0 or 1) rather than expressing appropriate uncertainty.</p></li><li><p><strong>Step Label Interference</strong>: Errors in verifying earlier steps can cascade, influencing the verification of later steps in the solution.</p></li><li><p><strong>Computational Overhead</strong>: Generating detailed verification chains requires more computation than simple discriminative judgments, though the performance benefits often justify this cost.</p></li></ol><h2><strong>The Broader Significance</strong></h2><p>ThinkPRM demonstrates a powerful principle: models can "think to verify" rather than simply "classify to verify." This represents a move toward more human-like verification systems that reason through solutions rather than making opaque judgments.The implications extend beyond academic research. As AI systems take on increasingly complex reasoning tasks in healthcare, scientific research, and critical infrastructure, reliable verification becomes essential. ThinkPRM's approach offers a path toward more trustworthy AI systems that can not only reason but also rigorously verify their reasoning processes.</p><h2><strong>Looking Forward</strong></h2><p>The ThinkPRM approach opens several exciting research directions:</p><ol><li><p><strong>Cross-domain verification</strong>: Further exploring how these models can generalize across different domains and types of reasoning tasks.</p></li><li><p><strong>Interactive verification</strong>: Developing systems that can ask clarifying questions when verification is uncertain.</p></li><li><p><strong>Self-correction</strong>: Using verification feedback to improve initial reasoning processes in a closed loop.</p></li><li><p><strong>Human-AI collaboration</strong>: Creating verification systems that can effectively collaborate with humans in complex reasoning tasks.</p></li></ol><h2><strong>Conclusion</strong></h2><p>ThinkPRM represents a significant advancement in AI verification technology, demonstrating that process reward models can achieve superior performance with dramatically less training data by leveraging generative, chain-of-thought reasoning. This "thinking verifier" approach aligns more closely with human verification practices and offers greater transparency into the verification process.</p><p>As AI systems tackle increasingly complex reasoning challenges, the ability to efficiently and reliably verify their work becomes ever more crucial. ThinkPRM shows that by teaching verification models to think through their judgments step by step, we can create more efficient, effective, and transparent verification systems &#8211; an essential step toward more trustworthy artificial intelligence.</p><div><hr></div><h3><strong>&#128077; If you enjoyed this article, give it a like and share it with your peers.</strong></h3><div><hr></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.llmwatch.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.llmwatch.com/subscribe?"><span>Subscribe now</span></a></p>]]></content:encoded></item><item><title><![CDATA[Multi-Agent Failure: What It Is and How to Prevent It]]></title><description><![CDATA[A Deep Dive into Failure Modes and System Design]]></description><link>https://www.llmwatch.com/p/multi-agent-failure-why-complex-ai</link><guid isPermaLink="false">https://www.llmwatch.com/p/multi-agent-failure-why-complex-ai</guid><dc:creator><![CDATA[Pascal Biese]]></dc:creator><pubDate>Tue, 29 Apr 2025 11:56:32 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/05525bc3-3095-4854-9007-c276f2b2d6fc_1536x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Multi-Agent Systems built using Large Language Models (LLMs) have emerged as a promising approach to complex problem-solving. By orchestrating multiple specialized agents working in concert, these systems aim to accomplish tasks that might be challenging for a single agent. However, despite growing enthusiasm and investment in these Multi-Agent LLM Systems (MAS), research reveals that their performance often falls short of expectations.</p><p>A <a href="https://arxiv.org/abs/2503.13657">recent study</a> presents a systematic analysis of failure patterns in MAS and introduces MAST (Multi-Agent System Failure Taxonomy), a comprehensive framework for understanding why these systems break down. In this deep dive, We'll explore the key findings from their research, what they tell us about building reliable Multi-Agent Systems, and where the field might go next.</p><h2><strong>The Problem: Promise vs. Reality</strong></h2><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Yqri!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F03b9e756-cf20-4767-beea-3e7acc753a23_634x517.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Yqri!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F03b9e756-cf20-4767-beea-3e7acc753a23_634x517.png 424w, https://substackcdn.com/image/fetch/$s_!Yqri!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F03b9e756-cf20-4767-beea-3e7acc753a23_634x517.png 848w, https://substackcdn.com/image/fetch/$s_!Yqri!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F03b9e756-cf20-4767-beea-3e7acc753a23_634x517.png 1272w, https://substackcdn.com/image/fetch/$s_!Yqri!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F03b9e756-cf20-4767-beea-3e7acc753a23_634x517.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Yqri!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F03b9e756-cf20-4767-beea-3e7acc753a23_634x517.png" width="634" height="517" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/03b9e756-cf20-4767-beea-3e7acc753a23_634x517.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:517,&quot;width&quot;:634,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:61939,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.llmwatch.com/i/162316430?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F03b9e756-cf20-4767-beea-3e7acc753a23_634x517.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Yqri!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F03b9e756-cf20-4767-beea-3e7acc753a23_634x517.png 424w, https://substackcdn.com/image/fetch/$s_!Yqri!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F03b9e756-cf20-4767-beea-3e7acc753a23_634x517.png 848w, https://substackcdn.com/image/fetch/$s_!Yqri!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F03b9e756-cf20-4767-beea-3e7acc753a23_634x517.png 1272w, https://substackcdn.com/image/fetch/$s_!Yqri!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F03b9e756-cf20-4767-beea-3e7acc753a23_634x517.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Despite their theoretical advantages, Multi-Agent Systems often disappoint in practice. The researchers found that even state-of-the-art MAS frameworks show high failure rates across popular benchmarks. For instance:</p><ul><li><p>ChatDev (ProgramDev): Models a miniature software company - agents take on design, coding and QA roles in sequence - but only gets about one-third of programming tasks right (33.3% correctness).</p></li><li><p>AppWorld (Test-C): Treats each everyday service (email, music, calendar, etc.) as its own agent under a supervisor orchestrator, yet fails 86.7% of its cross-app test cases.</p></li><li><p>HyperAgent (SWE-Bench Lite): Uses a central planner to hand off subtasks to navigator, editor and executor agents in a hierarchical software-engineering workflow, and still blows nearly three-quarters of its problems (74.7% failure).</p></li></ul><p>These sobering statistics raise a critical question: Why do systems with sophisticated architectures and powerful underlying LLMs fail so frequently?</p>
      <p>
          <a href="https://www.llmwatch.com/p/multi-agent-failure-why-complex-ai">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[d1: Scaling Reasoning in Diffusion Large Language Models]]></title><description><![CDATA[Learn what's great about Diffusion LLMs and how they are different from Transformers]]></description><link>https://www.llmwatch.com/p/d1-scaling-reasoning-in-diffusion</link><guid isPermaLink="false">https://www.llmwatch.com/p/d1-scaling-reasoning-in-diffusion</guid><dc:creator><![CDATA[Pascal Biese]]></dc:creator><pubDate>Tue, 22 Apr 2025 14:33:34 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/7bc0b1d8-211e-4c8c-b4d1-70c697f642fd_1536x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h3>TL;DR &#8212; Why <a href="https://arxiv.org/abs/2504.12216">this paper</a> matters</h3><p>Large language models that reason well are usually trained and fine&#8209;tuned in an <em>autoregressive</em> (left&#8209;to&#8209;right) way. <em>d1</em> shows, for the first time, that the same reinforcement&#8209;learning tricks that lifted autoregressive models to GPT&#8209;4&#8209;class reasoning can also lift <em>diffusion</em> language models, which generate text in a coarse&#8209;to&#8209;fine, non&#8209;sequential fashion. The authors introduce two key ingredients -<strong>masked supervised fine&#8209;tuning (SFT)</strong> and a new RL algorithm called <strong>diffu&#8209;GRPO</strong> -and demonstrate big gains on math and logic benchmarks without changing the base model size. </p><div><hr></div><h2>Reasoning beyond left&#8209;to&#8209;right text</h2><p>Most large language models we use today - GPT, Gemini, Claude - generate text one token after another in a single left&#8209;to&#8209;right pass. That <strong>autoregressive</strong> habit is intuitive for us humans, but it is not the only way to produce language. A fast&#8209;growing line of research borrows ideas from image diffusion models: start with a noisy, fully masked sentence and <em>denoise</em> it in several sweeps, filling in blanks until a coherent text emerges. The resulting <strong>diffusion LLMs (dLLMs)</strong>, such as LLaDA and Dream, can look at <em>future</em> context while writing the present token, often need fewer decoding steps, and open the door to parallel generation on hardware.</p><p>Yet all the spectacular reasoning progress you might have seen in DeepSeek&#8209;R1 or Kimi K1.5 was achieved with reinforcement learning (RL) algorithms - PPO, GRPO, variants of policy gradients - tailored to left&#8209;to&#8209;right models. Those methods rely on computing token&#8209;by&#8209;token log&#8209;probabilities. In dLLMs, the log&#8209;probability of a sentence is <em>not</em> factorized that way, so you cannot drop PPO into a diffusion model and hope it works.</p><p>That gap is exactly what <strong>d1 </strong>tries to close. The authors show that with just an 8&#8209;billion&#8209;parameter backbone and careful post&#8209;training, a diffusion model can rival or beat comparable autoregressive models on GSM8K math, MATH500 competition problems, mini&#8209;Sudoku, and the classic Countdown numbers game.</p><div><hr></div><div class="digest-post-embed" data-attrs="{&quot;nodeId&quot;:&quot;b1e1c51b-9727-4d71-b4b5-d47525ab1d57&quot;,&quot;caption&quot;:&quot;Welcome to State of AI Agents! Last week marked a decisive turning point for AI agents, with multiple tech giants unveiling ambitious agent strategies, frameworks, and deployments. This shift signals a fundamental transformation in how AI will be integrated into software and services going forward.&quot;,&quot;cta&quot;:null,&quot;showBylines&quot;:true,&quot;size&quot;:&quot;md&quot;,&quot;isEditorNode&quot;:true,&quot;title&quot;:&quot;State of AI Agents: Google Goes All-In on Agents&quot;,&quot;publishedBylines&quot;:[{&quot;id&quot;:48405812,&quot;name&quot;:&quot;Pascal Biese&quot;,&quot;bio&quot;:&quot;Human engineer reporting from the cutting edge of AI research.&quot;,&quot;photo_url&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/256e8620-1524-4496-a84d-7943a0edc098_512x512.png&quot;,&quot;is_guest&quot;:false,&quot;bestseller_tier&quot;:100}],&quot;post_date&quot;:&quot;2025-04-13T18:35:07.997Z&quot;,&quot;cover_image&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/7d05d9b8-ac4b-4c52-b87e-fe91ad976ddf_1610x1104.png&quot;,&quot;cover_image_alt&quot;:null,&quot;canonical_url&quot;:&quot;https://www.llmwatch.com/p/state-of-ai-agents-google-goes-all&quot;,&quot;section_name&quot;:&quot;State of AI Agents&quot;,&quot;video_upload_id&quot;:null,&quot;id&quot;:161195929,&quot;type&quot;:&quot;newsletter&quot;,&quot;reaction_count&quot;:17,&quot;comment_count&quot;:0,&quot;publication_id&quot;:null,&quot;publication_name&quot;:&quot;LLM Watch&quot;,&quot;publication_logo_url&quot;:&quot;https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5d95c476-43a7-4447-9081-9298a1fc325a_1280x1280.png&quot;,&quot;belowTheFold&quot;:false,&quot;youtube_url&quot;:null,&quot;show_links&quot;:null,&quot;feed_url&quot;:null}"></div><div><hr></div><h2>The d1 recipe at a glance</h2><p>The proposed pipeline has only two stages:</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!lbw1!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F859318f2-b038-4a4f-b1bc-50229d14fe06_803x229.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!lbw1!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F859318f2-b038-4a4f-b1bc-50229d14fe06_803x229.png 424w, https://substackcdn.com/image/fetch/$s_!lbw1!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F859318f2-b038-4a4f-b1bc-50229d14fe06_803x229.png 848w, https://substackcdn.com/image/fetch/$s_!lbw1!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F859318f2-b038-4a4f-b1bc-50229d14fe06_803x229.png 1272w, https://substackcdn.com/image/fetch/$s_!lbw1!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F859318f2-b038-4a4f-b1bc-50229d14fe06_803x229.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!lbw1!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F859318f2-b038-4a4f-b1bc-50229d14fe06_803x229.png" width="803" height="229" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/859318f2-b038-4a4f-b1bc-50229d14fe06_803x229.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:229,&quot;width&quot;:803,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:18316,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.llmwatch.com/i/161567535?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F859318f2-b038-4a4f-b1bc-50229d14fe06_803x229.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!lbw1!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F859318f2-b038-4a4f-b1bc-50229d14fe06_803x229.png 424w, https://substackcdn.com/image/fetch/$s_!lbw1!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F859318f2-b038-4a4f-b1bc-50229d14fe06_803x229.png 848w, https://substackcdn.com/image/fetch/$s_!lbw1!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F859318f2-b038-4a4f-b1bc-50229d14fe06_803x229.png 1272w, https://substackcdn.com/image/fetch/$s_!lbw1!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F859318f2-b038-4a4f-b1bc-50229d14fe06_803x229.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p>Why not merge the stages? In practice, SFT gives the model a basic &#8220;chain&#8209;of&#8209;thought grammar&#8221; - self&#8209;checks, backtracking, tidy XML tags - making the subsequent RL much more stable. The recipe is therefore called <strong>d1</strong>: &#8220;diffusion + one&#8209;two punch of SFT then RL.&#8221;</p>
      <p>
          <a href="https://www.llmwatch.com/p/d1-scaling-reasoning-in-diffusion">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[Augmented Work: The AI Teammates Are Coming]]></title><description><![CDATA[How AI and humans learn to work together over time]]></description><link>https://www.llmwatch.com/p/augmented-work-the-ai-teammates-are</link><guid isPermaLink="false">https://www.llmwatch.com/p/augmented-work-the-ai-teammates-are</guid><dc:creator><![CDATA[Pascal Biese]]></dc:creator><pubDate>Wed, 16 Apr 2025 16:17:31 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/a02e3668-77a8-435c-8a88-48a1f5c2766d_1536x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>The collaboration between humans and AI is no longer science fiction and will become an everyday reality sooner than later. From healthcare to creative design, AI systems are increasingly working alongside humans as teammates rather than mere tools. But how do these partnerships develop over time? How do they adapt to challenges? And what makes some human-AI teams thrive while others struggle?</p><p>Earlier today, I stumbled upon a new paper by <a href="https://arxiv.org/abs/2504.10918">Wang and colleagues</a> that&#8217;s trying to find answers to a lot of these questions. Through the lens of "human-agent teaming" (HAT), they are offering their perspective on how these collaborations might evolve. But rather than viewing HAT as a static arrangement, they approach it as a dynamic process that unfolds over time ("process dynamics perspective&#8221;). Let&#8217;s take a closer look ourselves.</p><h2><strong>Beyond the Tool Paradigm: AI as Teammate</strong></h2><p>For decades, our relationship with technology has largely followed a tool-user paradigm - humans wielded technology like a hammer or operated it like a vehicle. But today's AI systems, particularly those powered by large language models, possess unprecedented levels of autonomy, social capability, and proactiveness.</p><p>This has sparked interest in a new paradigm: human-agent teaming. In HAT, AI systems aren't just passive tools but active participants that share goals, distribute responsibilities, and engage in ongoing coordination with their human counterparts.</p><p>As the authors note: "Human-agent teaming (HAT) is defined as a collaborative framework in which humans and agents pursue shared goals, distribute responsibilities, and engage in ongoing coordination and negotiation to achieve joint objectives."</p><p>If one agrees with this perspective or not, current AI definitely <em>feels</em> different and this will ultimately change how we design and study human-AI interaction. Rather than focusing solely on usability or performance, researchers now need to consider team dynamics, shared understanding, and adaptive capacity.</p><p>From a practical point of view, we&#8217;ve only just begun to make the shift from static AI workflows to AI agents to Multi-Agent Systems. It&#8217;s already clear that AI (agent) orchestration will become a <em>very important</em> topic in the near future - but we often ignore the <em>human factor</em> when talking about these systems. I would argue, however, that humans will remain to play a key role in a lot of real-world settings, especially in high risk domains. So thinking about Multi-Agent Systems in isolation won&#8217;t be enough - and frameworks like HAT try to fill this gap.</p><h2><strong>The T4 Framework: A Lifecycle View of Human-AI Teams</strong></h2><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!jNwH!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F686208ba-4446-43b0-8f6b-c03484bd9142_1147x521.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!jNwH!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F686208ba-4446-43b0-8f6b-c03484bd9142_1147x521.png 424w, https://substackcdn.com/image/fetch/$s_!jNwH!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F686208ba-4446-43b0-8f6b-c03484bd9142_1147x521.png 848w, https://substackcdn.com/image/fetch/$s_!jNwH!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F686208ba-4446-43b0-8f6b-c03484bd9142_1147x521.png 1272w, https://substackcdn.com/image/fetch/$s_!jNwH!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F686208ba-4446-43b0-8f6b-c03484bd9142_1147x521.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!jNwH!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F686208ba-4446-43b0-8f6b-c03484bd9142_1147x521.png" width="1147" height="521" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/686208ba-4446-43b0-8f6b-c03484bd9142_1147x521.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:521,&quot;width&quot;:1147,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:193014,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.llmwatch.com/i/161466481?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F686208ba-4446-43b0-8f6b-c03484bd9142_1147x521.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!jNwH!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F686208ba-4446-43b0-8f6b-c03484bd9142_1147x521.png 424w, https://substackcdn.com/image/fetch/$s_!jNwH!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F686208ba-4446-43b0-8f6b-c03484bd9142_1147x521.png 848w, https://substackcdn.com/image/fetch/$s_!jNwH!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F686208ba-4446-43b0-8f6b-c03484bd9142_1147x521.png 1272w, https://substackcdn.com/image/fetch/$s_!jNwH!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F686208ba-4446-43b0-8f6b-c03484bd9142_1147x521.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">The T4 framework of HAT process dynamics. <a href="https://arxiv.org/abs/2504.10918">Source</a>.</figcaption></figure></div><p>One of the researchers&#8217; core contributions is the T4 framework &#8211; a comprehensive model that views HAT through four developmental phases:</p><ol><li><p><strong>Team Formation</strong>: Establishing team identity and shared goals</p></li><li><p><strong>Task and Role Development</strong>: Defining who does what and building individual competence</p></li><li><p><strong>Team Development</strong>: Creating shared understanding and effective coordination</p></li><li><p><strong>Team Improvement</strong>: Building adaptability and long-term viability</p></li></ol><p>What makes this framework powerful is how it integrates two key dynamics:</p><ul><li><p><strong>Task dynamics</strong>: The cyclical process through which team members set goals, execute tasks, evaluate outcomes, and adjust strategies</p></li><li><p><strong>Team developmental dynamics</strong>: The iterative progression through the four phases</p></li></ul><p>These dynamics don't operate in isolation but continuously influence each other. As the team completes tasks, it progresses through developmental phases. Meanwhile, each phase shapes how the team approaches its tasks.</p><p>The ultimate goal? A self-managing, self-regulating team capable of adapting to new challenges without external intervention.</p>
      <p>
          <a href="https://www.llmwatch.com/p/augmented-work-the-ai-teammates-are">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[DeepSeek-GRM: What It Is and Why You Should Care]]></title><description><![CDATA[A new generalist reward model from DeepSeek AI]]></description><link>https://www.llmwatch.com/p/deepseek-grm-what-it-is-and-why-you</link><guid isPermaLink="false">https://www.llmwatch.com/p/deepseek-grm-what-it-is-and-why-you</guid><dc:creator><![CDATA[Pascal Biese]]></dc:creator><pubDate>Tue, 08 Apr 2025 16:45:36 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/7712df91-9f91-4427-962c-3bcf88a1c095_1536x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>At the heart of recent AI advancements lies reward modeling&#8212;a mechanism for providing accurate reward signals to guide model optimization. A few days ago, a <a href="https://arxiv.org/abs/2504.02495v1">new paper</a> from DeepSeek AI introduced a novel approach that significantly advances the field by making reward models more effective and scalable across diverse domains. </p><p>While we've seen impressive progress in reward modeling for specific domains with clear verification criteria (like mathematics or coding), generating high-quality rewards for general domains remains challenging. </p><p>This might not be another &#8220;DeepSeek moment&#8221;, but it&#8217;s a pretty significant contribution that, for some reason, hasn&#8217;t really been picked up yet by the wider AI community&#8212;which is why I decided to write about it.</p><p>In this article, we'll explore the key innovations of their research, explain the technical details of the proposed method, Self-Principled Critique Tuning (SPCT) and inference-time scaling for reward models.</p><p>Before we dive into it though, let&#8217;s start with an (optional) executive summary that will give you a better idea of what DeepSeek-GRM is about.</p><div><hr></div><h2>Executive Summary</h2><p>DeepSeek AI&#8217;s latest release focuses on improving reward modeling (RM) for large language models (LLMs) through "inference-time scaling." In simple terms, it's about how to make reward models get better results by using more compute at inference time, rather than just making bigger models during training. </p><h3>Key themes and contributions:</h3><ol><li><p>Introducing Self-Principled Critique Tuning (SPCT), a method for generalist reward modeling</p></li><li><p>Developing a reward modeling approach using pointwise generative reward modeling (GRM)</p></li><li><p>Showing how to effectively scale inference computation for better reward quality</p></li><li><p>Creating DeepSeek-GRM models that outperform existing approaches</p></li><li><p>Introducing a meta RM to guide voting process for better scaling performance</p></li></ol><h3>Main challenges addressed:</h3><ol><li><p>Flexibility for different input types in reward modeling</p></li><li><p>Accurate reward generation in various domains</p></li><li><p>Scaling reward quality with increased inference compute</p></li><li><p>Learning scalable behaviors for better performance-compute scaling</p></li></ol><h3>Key methods:</h3><ol><li><p>Pointwise generative reward modeling (GRM)</p></li><li><p>Self-Principled Critique Tuning (SPCT) with:</p><ul><li><p>Rejective fine-tuning</p></li><li><p>Rule-based online reinforcement learning</p></li></ul></li><li><p>Inference-time scaling through:</p><ul><li><p>Parallel sampling</p></li><li><p>Meta RM-guided voting</p></li></ul></li></ol><h3>Key results:</h3><ol><li><p>DeepSeek-GRM outperforms baseline methods on various reward modeling benchmarks</p></li><li><p>Inference-time scaling provides significant performance improvements</p></li><li><p>Voting with meta RM guidance further boosts performance</p></li><li><p>Inference-time scaling can outperform training-time scaling (using larger models)</p></li></ol><div><hr></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.llmwatch.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">LLM Watch is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><div><hr></div><h2>The Challenge: Building Better Generalist Reward Models</h2><p>Reward modeling serves as the backbone of Reinforcement Learning from Human Feedback (RLHF), providing the signals that guide LLMs toward generating more helpful, harmless, and honest responses. While high-quality rewards exist for specific domains with clear verification criteria (like mathematics or coding), generating accurate rewards for general domains remains challenging. The researchers identify four key requirements for effective generalist reward models: </p><ol><li><p><strong>Flexibility</strong> to handle different input types (single responses, paired comparisons, or multiple candidates)</p></li><li><p><strong>Accuracy</strong> in generating rewards across diverse domains</p></li><li><p><strong>Inference-time scalability</strong> to improve reward quality with more compute</p></li><li><p><strong>Learning scalable behaviors</strong> that enhance performance as compute increases</p></li></ol><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!ZHCB!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3a129818-25da-4647-bc0d-f577d4ae484e_897x317.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!ZHCB!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3a129818-25da-4647-bc0d-f577d4ae484e_897x317.png 424w, https://substackcdn.com/image/fetch/$s_!ZHCB!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3a129818-25da-4647-bc0d-f577d4ae484e_897x317.png 848w, https://substackcdn.com/image/fetch/$s_!ZHCB!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3a129818-25da-4647-bc0d-f577d4ae484e_897x317.png 1272w, https://substackcdn.com/image/fetch/$s_!ZHCB!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3a129818-25da-4647-bc0d-f577d4ae484e_897x317.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!ZHCB!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3a129818-25da-4647-bc0d-f577d4ae484e_897x317.png" width="897" height="317" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/3a129818-25da-4647-bc0d-f577d4ae484e_897x317.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:317,&quot;width&quot;:897,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:36911,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.llmwatch.com/i/160838114?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3a129818-25da-4647-bc0d-f577d4ae484e_897x317.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!ZHCB!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3a129818-25da-4647-bc0d-f577d4ae484e_897x317.png 424w, https://substackcdn.com/image/fetch/$s_!ZHCB!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3a129818-25da-4647-bc0d-f577d4ae484e_897x317.png 848w, https://substackcdn.com/image/fetch/$s_!ZHCB!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3a129818-25da-4647-bc0d-f577d4ae484e_897x317.png 1272w, https://substackcdn.com/image/fetch/$s_!ZHCB!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3a129818-25da-4647-bc0d-f577d4ae484e_897x317.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Paradigms of reward modeling. <a href="https://arxiv.org/abs/2504.02495v1">Source</a>.</figcaption></figure></div><p>Existing approaches to reward modeling fall into several categories, each with limitations: </p><ul><li><p><strong>Scalar reward models</strong>: Output numerical scores but lack detailed reasoning</p></li><li><p><strong>Semi-scalar reward models</strong>: Generate both scores and critiques but struggle with scaling</p></li><li><p><strong>Generative reward models</strong>: Provide detailed textual critiques but need methods to scale effectively</p></li></ul><p>Additionally, reward models use either pointwise scoring (rating each response individually) or pairwise scoring (selecting the better of two responses). They found that pointwise generative reward models (GRMs) offer the best combination of flexibility and scalability potential. </p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!da1a!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8b164d61-ee1e-40d9-8a53-1ccbdcba6ca4_709x319.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!da1a!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8b164d61-ee1e-40d9-8a53-1ccbdcba6ca4_709x319.png 424w, https://substackcdn.com/image/fetch/$s_!da1a!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8b164d61-ee1e-40d9-8a53-1ccbdcba6ca4_709x319.png 848w, https://substackcdn.com/image/fetch/$s_!da1a!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8b164d61-ee1e-40d9-8a53-1ccbdcba6ca4_709x319.png 1272w, https://substackcdn.com/image/fetch/$s_!da1a!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8b164d61-ee1e-40d9-8a53-1ccbdcba6ca4_709x319.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!da1a!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8b164d61-ee1e-40d9-8a53-1ccbdcba6ca4_709x319.png" width="709" height="319" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8b164d61-ee1e-40d9-8a53-1ccbdcba6ca4_709x319.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:319,&quot;width&quot;:709,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:32319,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.llmwatch.com/i/160838114?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8b164d61-ee1e-40d9-8a53-1ccbdcba6ca4_709x319.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!da1a!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8b164d61-ee1e-40d9-8a53-1ccbdcba6ca4_709x319.png 424w, https://substackcdn.com/image/fetch/$s_!da1a!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8b164d61-ee1e-40d9-8a53-1ccbdcba6ca4_709x319.png 848w, https://substackcdn.com/image/fetch/$s_!da1a!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8b164d61-ee1e-40d9-8a53-1ccbdcba6ca4_709x319.png 1272w, https://substackcdn.com/image/fetch/$s_!da1a!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8b164d61-ee1e-40d9-8a53-1ccbdcba6ca4_709x319.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Scoring patterns. <a href="https://arxiv.org/abs/2504.02495v1">Source</a>.</figcaption></figure></div><h2>The Power of Principles in Reward Generation</h2><p>A key insight from the paper is that quality principles significantly improve reward generation. In preliminary experiments, the researchers discovered that properly selected principles boosted reward quality on benchmarks like Chat Hard and IFEval. Rather than using static principles, they explored allowing the model to dynamically generate principles tailored to each query and response pair. For example, when evaluating responses to a coding question, the model might generate principles like: </p><ol><li><p><strong>Code Correctness</strong> (Weight: 40%)</p></li><li><p><strong>Code Efficiency</strong> (Weight: 25%)</p></li><li><p><strong>Documentation Quality</strong> (Weight: 20%)</p></li><li><p><strong>Error Handling</strong> (Weight: 15%)</p></li></ol><p>These weighted principles create a structured framework for generating comprehensive critiques and more accurate rewards. </p><h2>Self-Principled Critique Tuning: Teaching Models to Generate Better Principles and Critiques</h2><p>The core contribution of this paper is Self-Principled Critique Tuning (SPCT), a method that enables generative reward models to learn generating adaptive, high-quality principles that guide critique generation. </p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!z6_3!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd1ab98a3-7c3d-4e81-8f63-e72484604730_1091x251.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!z6_3!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd1ab98a3-7c3d-4e81-8f63-e72484604730_1091x251.png 424w, https://substackcdn.com/image/fetch/$s_!z6_3!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd1ab98a3-7c3d-4e81-8f63-e72484604730_1091x251.png 848w, https://substackcdn.com/image/fetch/$s_!z6_3!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd1ab98a3-7c3d-4e81-8f63-e72484604730_1091x251.png 1272w, https://substackcdn.com/image/fetch/$s_!z6_3!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd1ab98a3-7c3d-4e81-8f63-e72484604730_1091x251.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!z6_3!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd1ab98a3-7c3d-4e81-8f63-e72484604730_1091x251.png" width="1091" height="251" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d1ab98a3-7c3d-4e81-8f63-e72484604730_1091x251.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:251,&quot;width&quot;:1091,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:99798,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.llmwatch.com/i/160838114?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd1ab98a3-7c3d-4e81-8f63-e72484604730_1091x251.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!z6_3!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd1ab98a3-7c3d-4e81-8f63-e72484604730_1091x251.png 424w, https://substackcdn.com/image/fetch/$s_!z6_3!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd1ab98a3-7c3d-4e81-8f63-e72484604730_1091x251.png 848w, https://substackcdn.com/image/fetch/$s_!z6_3!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd1ab98a3-7c3d-4e81-8f63-e72484604730_1091x251.png 1272w, https://substackcdn.com/image/fetch/$s_!z6_3!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd1ab98a3-7c3d-4e81-8f63-e72484604730_1091x251.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">A rather inaccesible visualization of SPCT training. <a href="https://arxiv.org/abs/2504.02495v1">Source</a>.</figcaption></figure></div><p>If this image is as confusing to you as it was to me: don&#8217;t worry, we&#8217;ll go through it together&#8212;step by step. </p><p>See the &#8220;RFT&#8221; and &#8220;RL&#8221; on the left? RFT stands for &#8220;Rejective Fine-Tuning&#8221; and RL for &#8220;Reinforcement Lerning&#8221;, or more specifically in this case, &#8220;Rule-Based Online Reinforcement Learning&#8221;. Those are the two phases of SPCT training. </p><h3>Phase 1: Rejective Fine-Tuning (RFT)</h3><p>The first phase prepares the model to generate well-formatted principles and critiques for various input types: </p><ol><li><p><strong>Data construction</strong>: Sampling trajectories (principles and critiques) from pretrained GRMs for queries and responses</p></li><li><p><strong>Rejection strategy</strong>: Discarding trajectories where predicted rewards don't align with ground truth, and removing "too easy" queries where all sampled trajectories are correct</p></li><li><p><strong>Optional hinting</strong>: For challenging cases, providing hints about the correct answer, though this sometimes leads to "shortcuts" in reasoning</p></li></ol><p>This phase creates a unified format for handling different numbers of responses, unlike previous approaches that used different formats for different scenarios. </p><h3>Phase 2: Rule-Based Online Reinforcement Learning (RL)</h3><p> The second phase refines the model's ability through online reinforcement learning: </p><ol><li><p>During rollout, the model generates principles and critiques for input queries and responses</p></li><li><p>The predicted rewards are extracted and compared to ground truth</p></li><li><p>The model receives binary rewards: +1 for correctly identifying the best response, -1 otherwise</p></li><li><p>The reinforcement learning objective includes a KL penalty to maintain format quality</p></li></ol><p>This approach encourages the model to develop principles that enable accurate discrimination between responses while avoiding severe biases. </p><h2>Scaling at Inference Time: More Compute, Better Results</h2><p>Once trained with SPCT, DeepSeek-GRM models demonstrate remarkable inference-time scalability through two key methods: </p><h3>1. Voting with Generated Rewards</h3><p> The basic approach involves: </p><ol><li><p>Running the model multiple times with different random seeds (parallel sampling)</p></li><li><p>Having each sample generate its own set of principles and critiques</p></li><li><p>Extracting numerical scores from each run</p></li><li><p>Summing the scores to determine the final ranking</p></li></ol><p>Since each sample generates different principles, the model evaluates responses from multiple perspectives, leading to more robust judgments. This process effectively expands the reward space, allowing for finer distinctions between responses. </p><h3>2. Meta Reward Modeling for Better Voting</h3><p> To further enhance performance, the researchers introduced a meta reward model that: </p><ol><li><p>Evaluates the quality of principles and critiques generated by the main reward model</p></li><li><p>Assigns scores to indicate which samples are more reliable</p></li><li><p>Guides the voting process by filtering out low-quality samples</p></li></ol><p>This meta RM is trained to identify the correctness of principles and critiques, using both positive and negative examples to improve performance. </p><h2>Impressive Results Across Diverse Benchmarks</h2><p>The researchers evaluated DeepSeek-GRM models on multiple reward modeling benchmarks including Reward Bench, PPE (both Preference and Correctness), RMB, and ReaLMistake. The results demonstrate several significant findings: </p><h3>DeepSeek-GRM Outperforms Baseline Methods</h3><p>Even without inference-time scaling, DeepSeek-GRM-27B (based on Gemma-2-27B) achieved superior overall performance compared to baseline methods like LLM-as-a-Judge and various scalar and semi-scalar reward models. Unlike scalar and semi-scalar reward models that showed strong biases (performing well on some benchmarks but poorly on others), DeepSeek-GRM demonstrated more consistent performance across diverse domains. </p><h3>Inference-Time Scaling Significantly Improves Performance</h3><p> The most striking results come from inference-time scaling: </p><ul><li><p>With 8 samples, DeepSeek-GRM-27B showed a performance increase of 2.7 percentage points</p></li><li><p>With meta RM guidance and 32 samples, it achieved a remarkable 4.9 percentage point improvement</p></li><li><p>This scaled performance reached levels competitive with or exceeding much larger models like GPT-4o and Nemotron-4-340B-Reward</p></li></ul><p> What's particularly noteworthy is that these improvements far exceeded those of baseline methods under similar scaling conditions. </p><h3>Inference-Time Scaling vs. Training-Time Scaling</h3><p> Perhaps most importantly, the researchers showed that inference-time scaling can be more effective than simply using larger models: </p><ul><li><p>DeepSeek-GRM-27B with meta RM-guided voting (32 samples) achieved performance comparable to a 671B parameter model using greedy decoding</p></li><li><p>This suggests that allocating more compute to inference might be more cost-effective than training larger models</p></li></ul><p> This finding challenges the conventional wisdom that "bigger is better" and offers a more nuanced view of compute allocation. </p><h2>Why SPCT Works: Technical Analysis</h2><p> The effectiveness of SPCT stems from several key design choices: </p><ol><li><p><strong>Unified pointwise GRM format</strong> enables flexible handling of different input types within the same model</p></li><li><p><strong>Self-generated principles</strong> create a tailored framework for critique that focuses on relevant criteria</p></li><li><p><strong>Rule-based reinforcement learning</strong> teaches the model to generate principles that enable effective discrimination</p></li><li><p><strong>Parallel sampling with diverse principles</strong> evaluates responses from multiple perspectives</p></li><li><p><strong>Quality control via meta RM</strong> ensures only reliable samples contribute to the final decision</p></li></ol><p>The principles serve as a form of chain-of-thought reasoning, helping the model organize its evaluation process and producing more consistent rewards. </p><h2>Limitations and Future Directions</h2><p> Despite its impressive performance, the approach has several limitations: </p><ol><li><p><strong>Efficiency challenges</strong>: Generative reward models require more computation than scalar models for each inference</p></li><li><p><strong>Domain-specific performance</strong>: DeepSeek-GRM still lags behind scalar models on some verifiable tasks</p></li><li><p><strong>Long-horizon reasoning</strong>: The researchers found that DeepSeek-R1, which uses extensive chain-of-thought reasoning, underperformed relative to its computational cost except on reasoning-intensive tasks</p></li></ol><p> These limitations point to several promising future directions: </p><ol><li><p><strong>Tool integration</strong>: Incorporating tools like code interpreters could enhance verification capabilities</p></li><li><p><strong>Efficiency improvements</strong>: Generating principles ahead of time could improve computational efficiency</p></li><li><p><strong>LLM evaluation applications</strong>: Using generated principles as interpretable criteria for identifying LLM weaknesses</p></li><li><p><strong>Co-scaling with policy models</strong>: Combining scalable rewards with scalable policies for even greater performance</p></li></ol><h2>Conclusion: A Significant Advance in Reward Modeling</h2><p>Self-Principled Critique Tuning represents a significant advance in reward modeling for LLMs. By enabling generative reward models to dynamically produce principles tailored to specific queries and responses, SPCT creates models that are both flexible across diverse inputs and effectively scalable with increased inference compute. </p><p>The impressive performance of DeepSeek-GRM models, particularly with inference-time scaling, suggests that this approach could become an important component of future LLM training pipelines. The finding that inference-time scaling can outperform training-time scaling (using larger models) is particularly noteworthy, offering a potentially more efficient path to improved reward modeling. </p><p>In a field where much attention focuses on scaling model size, this research offers a refreshing perspective: sometimes, using compute more intelligently at inference time can be more effective than simply building bigger models. </p><div><hr></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.llmwatch.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">LLM Watch is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><div><hr></div><div class="digest-post-embed" data-attrs="{&quot;nodeId&quot;:&quot;e948ad37-9162-4c71-890c-b524aaa1fed6&quot;,&quot;caption&quot;:&quot;Welcome to the first article of this new series! We&#8217;re one week into April there&#8217;s already a lot to talk about. From major product launches to strategic shifts to research insights: things aren&#8217;t slowing down, it seems.&quot;,&quot;cta&quot;:null,&quot;showBylines&quot;:true,&quot;size&quot;:&quot;md&quot;,&quot;isEditorNode&quot;:true,&quot;title&quot;:&quot;State of AI Agents: What OpenAI &amp; Google Are Planning&quot;,&quot;publishedBylines&quot;:[{&quot;id&quot;:48405812,&quot;name&quot;:&quot;Pascal Biese&quot;,&quot;bio&quot;:&quot;Human engineer reporting from the cutting edge of AI research.&quot;,&quot;photo_url&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/256e8620-1524-4496-a84d-7943a0edc098_512x512.png&quot;,&quot;is_guest&quot;:false,&quot;bestseller_tier&quot;:100}],&quot;post_date&quot;:&quot;2025-04-06T20:08:47.654Z&quot;,&quot;cover_image&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/961ab89c-bd33-4ab1-a178-f1471dce6b04_1280x720.png&quot;,&quot;cover_image_alt&quot;:null,&quot;canonical_url&quot;:&quot;https://www.llmwatch.com/p/state-of-ai-agents-what-openai-and&quot;,&quot;section_name&quot;:&quot;State of AI Agents&quot;,&quot;video_upload_id&quot;:null,&quot;id&quot;:160663962,&quot;type&quot;:&quot;newsletter&quot;,&quot;reaction_count&quot;:9,&quot;comment_count&quot;:0,&quot;publication_id&quot;:null,&quot;publication_name&quot;:&quot;LLM Watch&quot;,&quot;publication_logo_url&quot;:&quot;https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5d95c476-43a7-4447-9081-9298a1fc325a_1280x1280.png&quot;,&quot;belowTheFold&quot;:true,&quot;youtube_url&quot;:null,&quot;show_links&quot;:null,&quot;feed_url&quot;:null}"></div>]]></content:encoded></item><item><title><![CDATA[Think Like An AI Agent: Introduction to Planning with LLMs]]></title><description><![CDATA[Understand how AI agents plan with the help of Generative AI]]></description><link>https://www.llmwatch.com/p/think-like-an-ai-agent-introduction</link><guid isPermaLink="false">https://www.llmwatch.com/p/think-like-an-ai-agent-introduction</guid><dc:creator><![CDATA[Pascal Biese]]></dc:creator><pubDate>Wed, 12 Mar 2025 15:09:32 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!HyCn!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc67c4192-d9af-458e-a008-5212a1e55bb7_679x623.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Large Language Models (LLMs) are typically seen as the &#8220;brain&#8221; behind autonomous AI because of the crucial role they are playing in the current emergence of autonomous agents. However, while much attention has been paid to LLMs' reasoning and tool-learning capabilities, their planning abilities&#8212;crucial for effective agent autonomy&#8212;have received less systematic analysis. </p><p>A <a href="https://arxiv.org/abs/2402.02716">new paper</a> by Huang et al. provides  a comprehensive taxonomy of planning approaches in LLM-based agents. In this deep dive, we'll analyze the paper's taxonomy of approaches, examining implementation details, and discussing the challenges that lie ahead for researchers in this rapidly evolving field. </p><p>If you want to learn how LLMs are reshaping agentic planning, stay awhile and listen. </p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!HyCn!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc67c4192-d9af-458e-a008-5212a1e55bb7_679x623.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!HyCn!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc67c4192-d9af-458e-a008-5212a1e55bb7_679x623.png 424w, https://substackcdn.com/image/fetch/$s_!HyCn!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc67c4192-d9af-458e-a008-5212a1e55bb7_679x623.png 848w, https://substackcdn.com/image/fetch/$s_!HyCn!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc67c4192-d9af-458e-a008-5212a1e55bb7_679x623.png 1272w, https://substackcdn.com/image/fetch/$s_!HyCn!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc67c4192-d9af-458e-a008-5212a1e55bb7_679x623.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!HyCn!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc67c4192-d9af-458e-a008-5212a1e55bb7_679x623.png" width="679" height="623" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c67c4192-d9af-458e-a008-5212a1e55bb7_679x623.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:623,&quot;width&quot;:679,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:140882,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.llmwatch.com/i/158772910?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc67c4192-d9af-458e-a008-5212a1e55bb7_679x623.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!HyCn!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc67c4192-d9af-458e-a008-5212a1e55bb7_679x623.png 424w, https://substackcdn.com/image/fetch/$s_!HyCn!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc67c4192-d9af-458e-a008-5212a1e55bb7_679x623.png 848w, https://substackcdn.com/image/fetch/$s_!HyCn!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc67c4192-d9af-458e-a008-5212a1e55bb7_679x623.png 1272w, https://substackcdn.com/image/fetch/$s_!HyCn!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc67c4192-d9af-458e-a008-5212a1e55bb7_679x623.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Taxonomy on LLM-Agent planning. <a href="https://arxiv.org/abs/2402.02716">Source</a>.</figcaption></figure></div><h2>LLM Planning: The Fundamental Paradigm Shift</h2><p>Planning&#8212;the ability to generate a sequence of actions to achieve a goal&#8212;has traditionally been the domain of symbolic AI systems or reinforcement learning approaches. The paper defines the general planning formulation as:</p><div class="latex-rendered" data-attrs="{&quot;persistentExpression&quot;:&quot;p = (a0, a1, ..., at) = plan(E, g; &#920;, P)&quot;,&quot;id&quot;:&quot;BAFXIXQXNR&quot;}" data-component-name="LatexBlockToDOM"></div><p></p><p>Where p represents the plan (sequence of actions), E is the environment, g is the goal, &#920; represents the LLM parameters, and P the prompt.</p><p>Traditional planning methods face significant limitations:</p><ul><li><p>Symbolic methods require conversion of natural language into formal representations</p></li><li><p>These approaches lack error tolerance, failing with even minor errors</p></li><li><p>Reinforcement learning methods require extensive interaction data</p></li></ul><p>LLMs offer a promising alternative by leveraging their pre-trained knowledge and reasoning capabilities to approach planning in a more flexible and robust manner. </p><p>Huang et al. systematically categorize existing LLM-Agent planning approaches into five distinct but interconnected directions:</p><ol><li><p>Task Decomposition</p></li><li><p>Multi-Plan Selection</p></li><li><p>External Planner-Aided Planning</p></li><li><p>Reflection and Refinement</p></li><li><p>Memory-Augmented Planning</p></li></ol><p>Let's examine each approach in detail.</p>
      <p>
          <a href="https://www.llmwatch.com/p/think-like-an-ai-agent-introduction">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[Beyond Attention: Comparing Potential Transformer 2.0 Architectures]]></title><description><![CDATA[Google's Transformer 2.0 vs. Sakana AI's Transformer-Squared (Transformer&#178;)]]></description><link>https://www.llmwatch.com/p/beyond-attention-comparing-potential</link><guid isPermaLink="false">https://www.llmwatch.com/p/beyond-attention-comparing-potential</guid><dc:creator><![CDATA[Pascal Biese]]></dc:creator><pubDate>Tue, 04 Mar 2025 19:26:42 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/37e47a3b-4d74-4c09-b7ff-d97b38790ffe_1024x672.webp" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Let&#8217;s talk about Transformers today, and their potential future. While the original Transformer architecture revolutionized natural language processing with its self-attention mechanism, several recent architectural innovations address fundamental limitations that have restricted the practical application of these models: context length restrictions and task-specific adaptability.</p><p>Among those new architectures, there are two that are especially well-positioned to succeed the throne&#8212;one officially, the other unofficially:</p><ol><li><p><strong>Google's Titans</strong> (Transformer 2.0), which addresses the challenge of handling extremely long sequences and maintaining memory over time</p></li><li><p><strong>Sakana AI's Transformer&#178;</strong>, which focuses on creating dynamically adaptable models that can switch between tasks without extensive retraining</p></li></ol><p>These architectures represent complementary approaches to enhancing transformer capabilities. While Titans enhances memory to process sequences beyond traditional context windows, Transformer&#178; improves real-time adaptability to diverse tasks. Both address critical limitations in current large language models (LLMs) while taking fundamentally different architectural decisions.</p><p>Let's dive into the technical details of each system, analyze their strengths and potential applications, and consider how these approaches might shape the future of language models.</p><h2>Google's Titans: Extending Memory for Long Sequences</h2><h3>The Challenge of Context Length</h3><p>Traditional transformer models face a significant limitation in their ability to handle long sequences, stemming from the quadratic computational complexity of the self-attention mechanism. As sequence length increases, both memory requirements and computational costs grow proportionally to the square of the sequence length, making it impractical to process very long documents, videos, or time series data.</p><p>The Titans architecture addresses this fundamental challenge by introducing a new memory system that enables efficient processing of sequences up to 2 million tokens in length&#8212;far beyond the typical context windows of current state-of-the-art LLMs, which typically range from 8k to 128k tokens.</p><h3>Three-Tier Memory Architecture</h3><p>The key innovation in Titans is its three-tier memory system:</p><ol><li><p><strong>Short-term Memory</strong>: Implemented through the standard attention mechanism, this component handles immediate context similar to working memory in humans.</p></li><li><p><strong>Long-term Neural Memory</strong>: A specialized neural module designed to learn how to store and retrieve historical context over extended periods, enabling the model to maintain information well beyond the limitations of the attention window.</p></li><li><p><strong>Persistent Memory</strong>: A stable storage mechanism for task-specific knowledge that remains consistent throughout inference, providing a foundation for domain expertise.</p></li></ol><p>This represents a significant departure from traditional transformer design by incorporating explicit memory mechanisms inspired by human cognitive processes rather than relying solely on attention.</p>
      <p>
          <a href="https://www.llmwatch.com/p/beyond-attention-comparing-potential">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[Microsoft's Phi-4-Mini: Never Has Small Been This Good]]></title><description><![CDATA[Everything you need to know about their latest AI advancements]]></description><link>https://www.llmwatch.com/p/microsofts-phi-4-mini-never-has-small</link><guid isPermaLink="false">https://www.llmwatch.com/p/microsofts-phi-4-mini-never-has-small</guid><dc:creator><![CDATA[Pascal Biese]]></dc:creator><pubDate>Thu, 27 Feb 2025 14:35:00 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/e657c4df-980b-46c3-9b8b-e1466d61d86d_1280x720.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Creating efficient, powerful language models has been a significant challenge in AI research. Despite the headline-grabbing capabilities of massive LLMs with hundreds of billions of parameters, there's growing interest in developing smaller, more efficient models that maintain impressive capabilities while running on less hardware. Microsoft's latest entry in this space, Phi-4, represents a notable achievement in this direction. In this article, we'll dive deep into the technical innovations that make Phi-4 stand out&#8212;even among its fierce competition.</p><h2>The Phi Series: A Family of Compact Yet Powerful Models</h2><p>Microsoft's Phi series, and more specifically, its recently released <a href="https://arxiv.org/abs/2412.08905">fourth generation</a>, challenge the conventional wisdom that bigger is always better in language model development. These models demonstrate that with careful data curation, architectural optimizations, and strategic training approaches, even relatively small models can achieve impressive performance across language, vision, and speech tasks.</p><p>The <a href="https://huggingface.co/microsoft/Phi-4-multimodal-instruct/blob/main/phi_4_mm.tech_report.02252025.pdf">newest</a> Phi-4 release consists of two main models:</p><ul><li><p><strong>Phi-4-Mini</strong>: A 3.8-billion-parameter language model focused on text understanding and generation</p></li><li><p><strong>Phi-4-Multimodal</strong>: An extension that integrates vision and speech/audio capabilities while preserving the core language abilities</p></li></ul><p>What makes these models particularly interesting is their efficiency-to-performance ratio. Phi-4-Mini matches or exceeds models twice its size on certain tasks, especially in mathematics and coding, while Phi-4-Multimodal delivers competitive multimodal capabilities despite its modest parameter count.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!TUZt!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fef411aa9-9763-4373-9e83-614c89992167_1122x587.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!TUZt!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fef411aa9-9763-4373-9e83-614c89992167_1122x587.png 424w, https://substackcdn.com/image/fetch/$s_!TUZt!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fef411aa9-9763-4373-9e83-614c89992167_1122x587.png 848w, https://substackcdn.com/image/fetch/$s_!TUZt!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fef411aa9-9763-4373-9e83-614c89992167_1122x587.png 1272w, https://substackcdn.com/image/fetch/$s_!TUZt!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fef411aa9-9763-4373-9e83-614c89992167_1122x587.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!TUZt!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fef411aa9-9763-4373-9e83-614c89992167_1122x587.png" width="1122" height="587" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/ef411aa9-9763-4373-9e83-614c89992167_1122x587.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:587,&quot;width&quot;:1122,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:70675,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.llmwatch.com/i/158025380?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fef411aa9-9763-4373-9e83-614c89992167_1122x587.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!TUZt!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fef411aa9-9763-4373-9e83-614c89992167_1122x587.png 424w, https://substackcdn.com/image/fetch/$s_!TUZt!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fef411aa9-9763-4373-9e83-614c89992167_1122x587.png 848w, https://substackcdn.com/image/fetch/$s_!TUZt!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fef411aa9-9763-4373-9e83-614c89992167_1122x587.png 1272w, https://substackcdn.com/image/fetch/$s_!TUZt!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fef411aa9-9763-4373-9e83-614c89992167_1122x587.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Overview of the multimodal architecture of Phi-4-Multimodal. <a href="https://huggingface.co/microsoft/Phi-4-multimodal-instruct/blob/main/phi_4_mm.tech_report.02252025.pdf">Source</a>.</figcaption></figure></div><h2>Model Architecture: Technical Foundations</h2><h3>Language Model Architecture</h3><p>At its core, Phi-4-Mini employs a decoder-only Transformer architecture with several optimizations:</p><ul><li><p><strong>32 Transformer layers</strong> with hidden state size of 3,072</p></li><li><p><strong>Tied input/output embeddings</strong> to reduce memory consumption</p></li><li><p><strong>Expanded vocabulary</strong> of 200,064 tokens (up from Phi-3.5) to better support multilingual applications</p></li><li><p><strong>Group Query Attention (GQA)</strong> with 24 query heads and 8 key/value heads, reducing KV cache usage to one-third of standard size</p></li><li><p><strong>LongRoPE positional encoding</strong> to support 128K context length</p></li><li><p><strong>Fractional RoPE dimension</strong> ensuring 25% of attention head dimensions remain position-agnostic for smoother long-context handling</p></li></ul><p>The improved tokenizer deserves special mention as it enables better multilingual and multimodal processing. The expanded vocabulary provides more efficient token representation across languages, which is essential for applications beyond English.</p><h3>Multimodal Architecture: Mixture of LoRAs</h3><p>Perhaps the most innovative aspect of Phi-4-Multimodal is its "Mixture of LoRAs" approach for handling multiple modalities. This technique preserves the original language model capabilities while extending functionality to vision and speech:</p><ol><li><p><strong>Vision Modality</strong>:</p><ul><li><p>Image encoder based on SigLIP-400M (440M parameters)</p></li><li><p>Vision projector to align vision features with text embeddings</p></li><li><p>Vision adapter LoRA (370M parameters) for the language decoder</p></li></ul></li><li><p><strong>Speech/Audio Modality</strong>:</p><ul><li><p>Audio encoder with conformer blocks (460M parameters)</p></li><li><p>Audio projector to map speech features to text embedding space</p></li><li><p>Audio adapter LoRA (460M parameters)</p></li></ul></li></ol><p>The key trick here is that the base language model remains <strong>entirely frozen</strong> during multimodal training, with only the modality-specific components being updated. This prevents the interference and performance degradation that often occurs when fine-tuning language models for multimodal tasks.</p><h2>Training Methodology: The Secret Sauce</h2><p>The impressive performance of Phi-4-Mini doesn't come from architectural innovations alone&#8212;careful data curation and training strategies play a crucial role.</p>
      <p>
          <a href="https://www.llmwatch.com/p/microsofts-phi-4-mini-never-has-small">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[Introduction to LIMO: Less is More for LLM Reasoning]]></title><description><![CDATA[DeepSeek-R1 was only the beginning&#8212;find out how LIMO takes Data Efficiency to another level]]></description><link>https://www.llmwatch.com/p/introduction-to-limo-less-is-more</link><guid isPermaLink="false">https://www.llmwatch.com/p/introduction-to-limo-less-is-more</guid><dc:creator><![CDATA[Pascal Biese]]></dc:creator><pubDate>Wed, 12 Feb 2025 16:50:23 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!YqmK!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffae91f3a-7bec-4c6a-81d3-a3244f82b960_2641x1519.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>The recent paper <em>"LIMO: Less is More for Reasoning"</em> challenges foundational assumptions about how large language models (LLMs) acquire complex reasoning capabilities. By demonstrating that meticulously curated instruction data&#8212;as few as <strong>817 examples</strong>&#8212;can outperform models trained on 100x more data, the authors propose a paradigm shift in how we approach reasoning in LLMs. </p><p>This article explores the technical innovations, empirical findings, and broader implications of how LLM reasoning might change in the near future.</p><h3><strong>The LIMO Hypothesis: A New Lens on Reasoning</strong></h3><p>At its core, the <strong>Less-Is-More Reasoning (LIMO) Hypothesis </strong>states that:</p><blockquote><p><strong>Sophisticated reasoning capabilities emerge when two conditions converge</strong>:</p><ol><li><p><strong>Rich pre-trained knowledge</strong> embedded during LLM pre-training.</p></li><li><p><strong>Precisely orchestrated demonstrations</strong> (cognitive templates) during fine-tuning.</p></li></ol></blockquote><p>This hypothesis overturns the long-held belief that complex reasoning tasks require massive datasets (e.g., &gt;100k samples). Instead, LIMO argues that modern foundation models (e.g., Qwen2.5, Llama 3) already encode extensive domain knowledge; the challenge lies in <em>eliciting </em>this knowledge through high-quality examples that guide the model to "think" systematically.</p><p>They define the following requirements for success:</p><ul><li><p><strong>Pre-trained Knowledge</strong>: Modern LLMs like Qwen2.5-32B incorporate vast mathematical content during pre-training (3.7T tokens for Llama 3&#8217;s math-focused training).</p></li><li><p><strong>Inference-Time Scaling</strong>: Techniques allowing extended reasoning chains (e.g., parallel sampling, tree search) provide the "cognitive workspace" for multi-step problem-solving.</p></li></ul><div><hr></div><h3><strong>Data Curation: Quality Over Quantity</strong></h3><p>Data quality is core to the LIMO hypothesis and to ensure data quality, we need proper data curation. LIMO&#8217;s success hinges on its <strong>systematic data curation process</strong>, which prioritizes depth over breadth:</p><h4><strong>Question Selection</strong></h4><ul><li><p><strong>Difficulty</strong>: Problems are filtered using state-of-the-art models (e.g., Qwen2.5-Math-7B-Instruct), retaining only those with &lt;10% solve rates.</p></li><li><p><strong>Diversity</strong>: Selected from advanced benchmarks (AIME, OlympiadBench) and multilingual sources (Chinese Gaokao), ensuring coverage of algebra, geometry, and proof-based challenges.</p></li><li><p><strong>Out-of-Distribution (OOD) Focus</strong>: 30% of questions intentionally deviate from standard datasets to test generalization.</p></li></ul><p>From an initial pool of <strong>10M+ problems</strong>, only <strong>817 </strong>survived rigorous filtration.</p>
      <p>
          <a href="https://www.llmwatch.com/p/introduction-to-limo-less-is-more">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[DeepSeek-R1: What It Is & Why Everyone Is Talking About it]]></title><description><![CDATA[A gentle introduction to the open source GPT-o1 alternative]]></description><link>https://www.llmwatch.com/p/deepseek-r1-what-it-is-and-why-everyone</link><guid isPermaLink="false">https://www.llmwatch.com/p/deepseek-r1-what-it-is-and-why-everyone</guid><dc:creator><![CDATA[Pascal Biese]]></dc:creator><pubDate>Tue, 21 Jan 2025 18:02:00 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/d0eac864-e581-440e-9dca-77faa4ffe5f1_977x563.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>In recent years, Large Language Models (LLMs) have made significant advancements in their ability to understand and generate human-like text. These models, such as GPT-4 and Claude 3.5, have shown impressive performance in various natural language processing tasks. However, there is still room for improvement, particularly in the area of reasoning capabilities. To address this, researchers have explored a plethora of techniques &#8212; iteratively moving towards more and more complex data regimes and most recently, scaling up <a href="https://huggingface.co/spaces/HuggingFaceH4/blogpost-scaling-test-time-compute">test-time compute</a>.</p><p>Contrary to this trend, a <a href="https://www.deepseek.com/">rising AI research company</a> has now reported that they had more success with a far simpler approach: the use of reinforcement learning (RL) without having to rely on supervised fine-tuning <em>at all</em>.</p><h2>What is Reinforcement Learning? </h2><p>Before diving into DeepSeek's approach, let's first understand what reinforcement learning is. Imagine you're teaching a dog a new trick. You give the dog a treat every time it performs the desired action, such as sitting or rolling over. The dog learns to associate the action with the reward and becomes more likely to repeat the behavior in the future. This is the basic principle behind reinforcement learning.</p><p>In the context of LLMs, the model is the "dog," and the reward is a score that measures how well the model performs on a specific task. The model learns to generate text that maximizes the reward, thereby improving its performance on the task.</p><h3>DeepSeek-R1-Zero: Pure Reinforcement Learning </h3><p>The first step in DeepSeek's approach was to apply RL directly to the base model, DeepSeek-V3-Base, without any supervised fine-tuning (SFT). This model, called DeepSeek-R1-Zero, was allowed to explore different reasoning strategies, such as Chain-of-Thought (CoT), to solve complex problems.</p><p>Think of CoT as a step-by-step thought process that the model goes through to arrive at a solution. For example, if the model is asked, "What is the capital of France?", it might generate the following CoT:</p><ol><li><p>France is a country in Europe.</p></li><li><p>The capital of a country is usually its largest and most important city.</p></li><li><p>Paris is the largest and most important city in France.</p></li><li><p>Therefore, the capital of France is Paris.</p></li></ol><p>By exploring different CoT strategies, DeepSeek-R1-Zero was able to develop powerful reasoning capabilities without any supervised data. This was a significant finding, as it demonstrated that LLMs could improve their reasoning abilities through pure RL.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!j7YB!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F01979ee5-fcf6-435f-8f92-f03dd535f749_977x563.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!j7YB!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F01979ee5-fcf6-435f-8f92-f03dd535f749_977x563.png 424w, https://substackcdn.com/image/fetch/$s_!j7YB!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F01979ee5-fcf6-435f-8f92-f03dd535f749_977x563.png 848w, https://substackcdn.com/image/fetch/$s_!j7YB!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F01979ee5-fcf6-435f-8f92-f03dd535f749_977x563.png 1272w, https://substackcdn.com/image/fetch/$s_!j7YB!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F01979ee5-fcf6-435f-8f92-f03dd535f749_977x563.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!j7YB!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F01979ee5-fcf6-435f-8f92-f03dd535f749_977x563.png" width="977" height="563" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/01979ee5-fcf6-435f-8f92-f03dd535f749_977x563.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:563,&quot;width&quot;:977,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:126222,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!j7YB!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F01979ee5-fcf6-435f-8f92-f03dd535f749_977x563.png 424w, https://substackcdn.com/image/fetch/$s_!j7YB!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F01979ee5-fcf6-435f-8f92-f03dd535f749_977x563.png 848w, https://substackcdn.com/image/fetch/$s_!j7YB!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F01979ee5-fcf6-435f-8f92-f03dd535f749_977x563.png 1272w, https://substackcdn.com/image/fetch/$s_!j7YB!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F01979ee5-fcf6-435f-8f92-f03dd535f749_977x563.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">DeepSeek-R1-Zero having an &#8220;aha&#8221; moment. <a href="https://github.com/deepseek-ai/DeepSeek-V3">Source</a>.</figcaption></figure></div>
      <p>
          <a href="https://www.llmwatch.com/p/deepseek-r1-what-it-is-and-why-everyone">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[Stream of Search: Teaching Language Models the Language of Search ]]></title><description><![CDATA[Generative AI needs to learn from mistakes]]></description><link>https://www.llmwatch.com/p/stream-of-search-teaching-language</link><guid isPermaLink="false">https://www.llmwatch.com/p/stream-of-search-teaching-language</guid><dc:creator><![CDATA[Pascal Biese]]></dc:creator><pubDate>Wed, 30 Oct 2024 18:01:25 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!Awzc!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe964225-47c1-4fec-a6bc-081e6af71b4c_860x300.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Language models have demonstrated remarkable abilities in various tasks, from natural language processing to code generation. However, these models often struggle with complex problem-solving that requires planning, searching, and backtracking. A recent paper by Gandhi et al. (2024) introduces a novel approach called Stream of Search (SoS), which aims to teach language models to solve problems by searching in language, without relying on external components.</p><p>The Stream of Search framework systematizes the elements of search into a unified language, enabling the representation of diverse search strategies in a common format. By training language models on these "streams of search," the authors demonstrate that the models can learn to solve problems more effectively than those trained solely on optimal solution trajectories. Furthermore, the SoS models can self-improve by optimizing for correctness using reinforcement learning techniques such as STaR and APA.</p><p>In this article, we will delve into the Stream of Search framework, exploring its key components, the problem setup, and the experimental results that showcase its potential for enhancing the problem-solving capabilities of language models.</p><h2>The Language of Search </h2><p>At the core of the Stream of Search framework is a vocabulary of primitive operations that define the components of various search algorithms. </p><p>These operations include:</p><ol><li><p>Current State (sc): The state being explored.</p></li><li><p>Goal State (sg): The target state.</p></li><li><p>State Queue (Sq): The states at the frontier of the trajectory that haven't been explored yet.</p></li><li><p>State Expansion Function (SE): A function that explores a state adjacent to the current state based on a transition function.</p></li><li><p>Exploration Choice: Choosing the order of states to explore (e.g., breadth-first search or depth-first search).</p></li><li><p>Pruning: Discarding states or subtrees that are unlikely to lead to a solution.</p></li><li><p>Backtracking: Moving between explored nodes to choose the next state for expansion.</p></li><li><p>Goal Check: Checking if the current state is the goal state.</p></li><li><p>Heuristic (h): A function that approximates the distance of the current state from the goal state, guiding the search process.</p></li></ol><p>By representing these operations in language, the Stream of Search framework allows for the creation of a dataset with diverse search strategies. Some operations, such as the current state, goal state, backtracking, goal checks, and exploration choices, are explicitly represented in the search trajectory. Others, like heuristic functions, state values, and pruning strategies, are kept implicit, encouraging the model to internalize abstract representations that can be improved through training.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Awzc!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe964225-47c1-4fec-a6bc-081e6af71b4c_860x300.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Awzc!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe964225-47c1-4fec-a6bc-081e6af71b4c_860x300.png 424w, https://substackcdn.com/image/fetch/$s_!Awzc!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe964225-47c1-4fec-a6bc-081e6af71b4c_860x300.png 848w, https://substackcdn.com/image/fetch/$s_!Awzc!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe964225-47c1-4fec-a6bc-081e6af71b4c_860x300.png 1272w, https://substackcdn.com/image/fetch/$s_!Awzc!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe964225-47c1-4fec-a6bc-081e6af71b4c_860x300.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Awzc!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe964225-47c1-4fec-a6bc-081e6af71b4c_860x300.png" width="860" height="300" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/fe964225-47c1-4fec-a6bc-081e6af71b4c_860x300.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:300,&quot;width&quot;:860,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:76014,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Awzc!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe964225-47c1-4fec-a6bc-081e6af71b4c_860x300.png 424w, https://substackcdn.com/image/fetch/$s_!Awzc!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe964225-47c1-4fec-a6bc-081e6af71b4c_860x300.png 848w, https://substackcdn.com/image/fetch/$s_!Awzc!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe964225-47c1-4fec-a6bc-081e6af71b4c_860x300.png 1272w, https://substackcdn.com/image/fetch/$s_!Awzc!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffe964225-47c1-4fec-a6bc-081e6af71b4c_860x300.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Visualization of how a search process is translated into a stream of search. <a href="https://arxiv.org/abs/2404.03683">Source</a>.</figcaption></figure></div><h2>Problem Setup: The Game of Countdown </h2><p>To demonstrate the utility of the Stream of Search framework, the authors focus on a generalization of the 24 Game called Countdown. In this game, a set of input numbers must be combined using simple arithmetic operations to reach a target number. Countdown presents a challenging search problem due to its high branching factor, requiring planning, search, and backtracking to solve.</p><p>The authors construct a synthetic dataset of 500,000 search trajectories using a set of diverse and suboptimal symbolic search strategies based on breadth-first search (BFS) and depth-first search (DFS) with two simple heuristic functions. The heuristics used are:</p><ol><li><p>The absolute difference between the sum of the remaining options and the target.</p></li><li><p>The distance to the factors of the target.</p></li></ol><p>The search trajectories are serialized as strings, representing a list of tree nodes or states in the order of traversal. The dataset is then used to train language models in two conditions:</p><ol><li><p>Optimal Paths (OP): The model is trained to predict the correct and optimal path for all problems in the dataset.</p></li><li><p>Stream of Search (SoS): The model is trained on search trajectories sampled from different search strategies.</p></li></ol><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!JnEc!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F75ce4c63-66ed-451f-bae8-baec6fdb7ea3_823x511.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!JnEc!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F75ce4c63-66ed-451f-bae8-baec6fdb7ea3_823x511.png 424w, https://substackcdn.com/image/fetch/$s_!JnEc!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F75ce4c63-66ed-451f-bae8-baec6fdb7ea3_823x511.png 848w, https://substackcdn.com/image/fetch/$s_!JnEc!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F75ce4c63-66ed-451f-bae8-baec6fdb7ea3_823x511.png 1272w, https://substackcdn.com/image/fetch/$s_!JnEc!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F75ce4c63-66ed-451f-bae8-baec6fdb7ea3_823x511.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!JnEc!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F75ce4c63-66ed-451f-bae8-baec6fdb7ea3_823x511.png" width="823" height="511" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/75ce4c63-66ed-451f-bae8-baec6fdb7ea3_823x511.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:511,&quot;width&quot;:823,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:149300,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!JnEc!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F75ce4c63-66ed-451f-bae8-baec6fdb7ea3_823x511.png 424w, https://substackcdn.com/image/fetch/$s_!JnEc!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F75ce4c63-66ed-451f-bae8-baec6fdb7ea3_823x511.png 848w, https://substackcdn.com/image/fetch/$s_!JnEc!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F75ce4c63-66ed-451f-bae8-baec6fdb7ea3_823x511.png 1272w, https://substackcdn.com/image/fetch/$s_!JnEc!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F75ce4c63-66ed-451f-bae8-baec6fdb7ea3_823x511.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Overview of the Stream of Search (SoS) framework. <a href="https://arxiv.org/abs/2404.03683">Source</a>.</figcaption></figure></div><h2>Experimental Results </h2><p>The authors train a GPT-Neo model with 250M parameters on the Countdown dataset and evaluate its performance on held-out problems. The results demonstrate that the model trained on streams of search (SoS) outperforms the model trained on optimal solutions (OP). The SoS model achieves an accuracy of 51.27% on held-out inputs, compared to 25.73% for the OP model. This finding highlights the importance of exposing models to the messy process of problem-solving, including exploration and backtracking, rather than only the ideal solution steps.</p><p>To understand the strategies employed by the trained SoS model, the authors measure the alignment of the model-generated search trajectories with symbolic strategies. They find that the SoS model does not predominantly use any single strategy from its training data but instead exhibits a higher correlation with strategies that use the sum heuristic.</p><div><hr></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.llmwatch.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">LLM Watch is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><div><hr></div><h3>Policy Improvement with Stream of Search </h3><p>The authors further investigate whether the SoS model can self-improve with feedback based on correctness and efficiency. They employ two reinforcement learning strategies: expert iteration using STaR (Self-Taught Reasoner) and Advantage-Induced Policy Alignment (APA).</p><p>In the STaR approach, the authors use problems from the training dataset to generate 100,000 correct trajectories. These trajectories are then used to fine-tune the model iteratively until convergence in performance on the validation set is observed. After three iterations of STaR fine-tuning, the SoS+STaR model solves an additional 5% of the held-out inputs test set compared to the base SoS model.</p><p>Alternatively, the authors use APA, an Actor-Critic reinforcement learning technique that involves creating a copy of the language model to serve as a value network, which is then used to enhance the policy (the original language model). A straightforward reward function is defined, considering the correctness and length of the generated trajectory. The authors observe that updating the reference policy whenever the validation reward converges results in further policy improvement. After fine-tuning with APA, the SoS model achieves an improvement of about 6% over the base SoS model.</p><p>Analysis of the fine-tuned models reveals that both the STaR and APA models visit more states associated with the 'multiply' heuristic, which measures distance to the factors of the target. The APA model, in particular, diverges more from the symbolic strategies, indicating that it employs different strategies for searching and potentially discovers novel heuristics and search methods.</p><p>To further evaluate the performance of the improved models, the authors select 10,000 problems from the SoS training set that were unsolved by symbolic strategies and 10,000 difficult problems that none of the symbolic strategies used to train the SoS models can solve. Remarkably, the models solve approximately 36% of the previously unsolved problems and about 4% of the difficult problems.</p><h2>Discussion and Future Directions </h2><p>The SoS framework introduces a new approach to teaching language models to solve problems by searching in language, without relying on external components. By systematizing the elements of search into a unified language, the authors demonstrate that training language models on diverse streams of search leads to superior performance compared to models trained solely on optimal trajectories.</p><p>This addresses criticisms of language models for planning and problem-solving, such as the snowballing of errors and difficulty in lookahead tasks. By teaching models to backtrack and explore alternative paths, SoS enables language models to consider multiple possible outcomes before committing to a course of action. Crucially, SoS leads language models to learn an internal 'world model' for search, allowing for more adaptable and generalizable search compared to symbolic search that relies on an explicit environment model.</p><p>While the empirical results in the paper are restricted to the game of Countdown, the authors are optimistic that the SoS framework can be extended to more challenging, real-world tasks. Future research could explore integrating subgoals and hierarchical planning, as well as incorporating reflection and self-evaluation to enable models to discover and improve novel search strategies.</p><p>Generating the initial SoS dataset can be challenging, as it is not always feasible to create symbolic search algorithms to solve problems. An important question for future research is how well search abilities transfer between domains and between formal and informal domains.</p><h2>Conclusion </h2><p>SoS is a promising step forward in teaching language models to solve problems through structured search with backtracking, heuristic state evaluation, and world modeling. By exposing language models to diverse search strategies and iteratively refining them, the approach unlocks the potential of these models to tackle complex problems and discover new ways to solve them.</p><p>Frameworks like this will play a crucial role in enhancing the problem-solving capabilities of language models. By embracing the messy process of exploration and backtracking, and learning from productive mistakes, language models can develop more robust and adaptable search strategies, paving the way for their application in a wide range of real-world problem-solving scenarios.</p><p>AI that isn&#8217;t able to explore - and learn from mistakes - will always be limited by design. Unless we find a way to create flawless and complete training data for everything that we want it to do in this world - which, of course, is extremely unlikely.</p><div><hr></div><h3>&#128077; If you enjoyed this article, give it a like and share it with your peers.</h3><p>And in case you want to continue reading, here&#8217;s my previous research summary on StructRAG, a framework that combines the best of both worlds - graph-based and standard RAG:</p><div class="digest-post-embed" data-attrs="{&quot;nodeId&quot;:&quot;77b9de3d-6fe5-4b5d-8552-737d868cf80a&quot;,&quot;caption&quot;:&quot;Effectively leveraging external knowledge to enhance the performance of large language models (LLMs) on knowledge-based tasks has become a crucial challenge. Retrieval-augmented generation (RAG) methods have emerged as a promising solution, providing LLMs with relevant information from external sources to improve their factual accuracy and reasoning capabilities. However, existing RAG approaches struggle with knowledge-intensive reasoning tasks, where the required information is often scattered across multiple documents, making it difficult to accurately identify and integrate key pieces of information for global reasoning.&quot;,&quot;cta&quot;:null,&quot;showBylines&quot;:true,&quot;size&quot;:&quot;sm&quot;,&quot;isEditorNode&quot;:true,&quot;title&quot;:&quot;StructRAG: Succeeding Where Graph RAG Fails&quot;,&quot;publishedBylines&quot;:[{&quot;id&quot;:48405812,&quot;name&quot;:&quot;Pascal Biese&quot;,&quot;bio&quot;:&quot;Data Scientist &amp; ML Engineer writing about Large Language Models (LLMs), Natural Language Processing (NLP) and Artificial Intelligence&quot;,&quot;photo_url&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/256e8620-1524-4496-a84d-7943a0edc098_512x512.png&quot;,&quot;is_guest&quot;:false,&quot;bestseller_tier&quot;:null}],&quot;post_date&quot;:&quot;2024-10-15T18:01:06.687Z&quot;,&quot;cover_image&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e992656c-11f0-4fcf-b681-b13a942efac2_503x365.png&quot;,&quot;cover_image_alt&quot;:null,&quot;canonical_url&quot;:&quot;https://www.llmwatch.com/p/structrag-succeeding-where-graph&quot;,&quot;section_name&quot;:&quot;Research Summaries&quot;,&quot;video_upload_id&quot;:null,&quot;id&quot;:150204746,&quot;type&quot;:&quot;newsletter&quot;,&quot;reaction_count&quot;:46,&quot;comment_count&quot;:0,&quot;publication_id&quot;:null,&quot;publication_name&quot;:&quot;LLM Watch&quot;,&quot;publication_logo_url&quot;:&quot;https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5d95c476-43a7-4447-9081-9298a1fc325a_1280x1280.png&quot;,&quot;belowTheFold&quot;:true,&quot;youtube_url&quot;:null,&quot;show_links&quot;:null,&quot;feed_url&quot;:null}"></div><div><hr></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.llmwatch.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">LLM Watch is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[AI for Science: How to turn your LLM into an Innovation Machine]]></title><description><![CDATA[The Power of Iterative Planning and Search in LLM-Based Scientific Innovation]]></description><link>https://www.llmwatch.com/p/ai-for-science-how-to-turn-your-llm</link><guid isPermaLink="false">https://www.llmwatch.com/p/ai-for-science-how-to-turn-your-llm</guid><dc:creator><![CDATA[Pascal Biese]]></dc:creator><pubDate>Tue, 22 Oct 2024 15:01:15 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!8i1I!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F88df1880-ba40-4f24-9ebc-227827251bb4_1242x604.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Large Language Models (LLMs) have demonstrated remarkable capabilities across various domains, including scientific innovation. By harnessing the power of these models, researchers aim to accelerate the discovery process and generate novel research ideas. However, existing LLM-based methods often struggle to produce truly diverse and innovative concepts due to their limited ability to acquire and integrate external knowledge effectively. To address this challenge, a new approach called Nova has been introduced, which combines iterative planning and search to enhance the creative potential of LLM-based systems.</p><h2>The Nova Pipeline: A Three-Stage Approach to Scientific Innovation</h2><p>The Nova pipeline streamlines the research process through three key stages: initial idea generation, iterative refinement, and detailed completion. This systematic approach ensures that the generated ideas are not only novel but also well-developed and feasible.</p><div><hr></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.llmwatch.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">LLM Watch is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><div><hr></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!8i1I!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F88df1880-ba40-4f24-9ebc-227827251bb4_1242x604.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!8i1I!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F88df1880-ba40-4f24-9ebc-227827251bb4_1242x604.png 424w, https://substackcdn.com/image/fetch/$s_!8i1I!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F88df1880-ba40-4f24-9ebc-227827251bb4_1242x604.png 848w, https://substackcdn.com/image/fetch/$s_!8i1I!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F88df1880-ba40-4f24-9ebc-227827251bb4_1242x604.png 1272w, https://substackcdn.com/image/fetch/$s_!8i1I!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F88df1880-ba40-4f24-9ebc-227827251bb4_1242x604.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!8i1I!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F88df1880-ba40-4f24-9ebc-227827251bb4_1242x604.png" width="1242" height="604" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/88df1880-ba40-4f24-9ebc-227827251bb4_1242x604.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:604,&quot;width&quot;:1242,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:349762,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!8i1I!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F88df1880-ba40-4f24-9ebc-227827251bb4_1242x604.png 424w, https://substackcdn.com/image/fetch/$s_!8i1I!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F88df1880-ba40-4f24-9ebc-227827251bb4_1242x604.png 848w, https://substackcdn.com/image/fetch/$s_!8i1I!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F88df1880-ba40-4f24-9ebc-227827251bb4_1242x604.png 1272w, https://substackcdn.com/image/fetch/$s_!8i1I!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F88df1880-ba40-4f24-9ebc-227827251bb4_1242x604.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Nova Pipeline. The Pipeline includes initial seed idea generation, seed idea iteration, and idea completion. <a href="https://arxiv.org/abs/2410.14255">Source</a>.</figcaption></figure></div><h3>Stage 1: Initial Seed Idea Generation</h3><p>The first stage of the Nova pipeline focuses on generating diverse and novel seed ideas based on an input paper. To achieve this, the system employs a multi-source seed idea generation module that leverages the LLM's internal knowledge, related literature, and scientific discovery techniques.</p><p>One of the key components of this module is the knowledge tracking system, which addresses the shortcomings of previous approaches by monitoring the latest publications in the field. By identifying influential recent papers based on user engagement metrics across various platforms, such as social media, forums, and GitHub, Nova ensures that the generated ideas are informed by the most current insights.</p><p>To further increase the diversity of the generated ideas, Nova utilizes 10 fundamental scientific discovery methods derived from <a href="https://en.wikipedia.org/wiki/Paradigm_shift">Kuhn's paradigm</a> of scientific discovery. These methods help identify new research problems by analyzing anomalies in existing approaches, exploring theoretical boundaries, and integrating interdisciplinary knowledge.</p><p>Additionally, Nova employs self-correction mechanics, such as self-check, self-critique, and reflection, to prevent hallucination and improve the logicality of the generated seed ideas. By the end of this stage, the system generates 15 seed ideas for each input paper.</p>
      <p>
          <a href="https://www.llmwatch.com/p/ai-for-science-how-to-turn-your-llm">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[StructRAG: Succeeding Where Graph RAG Fails]]></title><description><![CDATA[Enhancing Knowledge-Intensive Reasoning in LLMs through Information Structurization]]></description><link>https://www.llmwatch.com/p/structrag-succeeding-where-graph</link><guid isPermaLink="false">https://www.llmwatch.com/p/structrag-succeeding-where-graph</guid><dc:creator><![CDATA[Pascal Biese]]></dc:creator><pubDate>Tue, 15 Oct 2024 18:01:06 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/e992656c-11f0-4fcf-b681-b13a942efac2_503x365.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Effectively leveraging external knowledge to enhance the performance of large language models (LLMs) on knowledge-based tasks has become a crucial challenge. Retrieval-augmented generation (RAG) methods have emerged as a promising solution, providing LLMs with relevant information from external sources to improve their factual accuracy and reasoning capabilities. However, existing RAG approaches struggle with knowledge-intensive reasoning tasks, where the required information is often scattered across multiple documents, making it difficult to accurately identify and integrate key pieces of information for global reasoning.</p><p>Inspired by cognitive theories suggesting that humans convert raw information into structured knowledge when tackling complex reasoning tasks, researchers from the Chinese Academy of Sciences and Alibaba Group have proposed a new framework called StructRAG. This framework introduces a hybrid information structurization mechanism that constructs and utilizes structured knowledge in the most suitable format based on the specific requirements of the task at hand. By mimicking human-like thinking processes, StructRAG aims to enhance the performance of LLMs on knowledge-intensive reasoning tasks.</p><p>In this article, we will delve into the details of the StructRAG framework, exploring its key components and how they work together to improve RAG performance. We will also discuss the training process of the hybrid structure router, a critical module in StructRAG, and present the experimental results demonstrating the effectiveness of this approach.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!tuFC!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8418326-eb1b-4f86-90c9-458e846840fd_990x316.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!tuFC!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8418326-eb1b-4f86-90c9-458e846840fd_990x316.png 424w, https://substackcdn.com/image/fetch/$s_!tuFC!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8418326-eb1b-4f86-90c9-458e846840fd_990x316.png 848w, https://substackcdn.com/image/fetch/$s_!tuFC!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8418326-eb1b-4f86-90c9-458e846840fd_990x316.png 1272w, https://substackcdn.com/image/fetch/$s_!tuFC!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8418326-eb1b-4f86-90c9-458e846840fd_990x316.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!tuFC!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8418326-eb1b-4f86-90c9-458e846840fd_990x316.png" width="990" height="316" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d8418326-eb1b-4f86-90c9-458e846840fd_990x316.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:316,&quot;width&quot;:990,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:112260,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!tuFC!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8418326-eb1b-4f86-90c9-458e846840fd_990x316.png 424w, https://substackcdn.com/image/fetch/$s_!tuFC!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8418326-eb1b-4f86-90c9-458e846840fd_990x316.png 848w, https://substackcdn.com/image/fetch/$s_!tuFC!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8418326-eb1b-4f86-90c9-458e846840fd_990x316.png 1272w, https://substackcdn.com/image/fetch/$s_!tuFC!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8418326-eb1b-4f86-90c9-458e846840fd_990x316.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Overview of the StructRAG framework. <a href="https://arxiv.org/abs/2410.08815">Source.</a></figcaption></figure></div><h2>The Framework</h2><p>The StructRAG framework consists of three main modules that work sequentially to identify the optimal structure type, construct structured knowledge, and utilize that knowledge for accurate reasoning. Let's take a closer look at each of these modules.</p><ol><li><p>Hybrid Structure Router: The hybrid structure router is the core component of StructRAG, responsible for determining the most appropriate structure type for a given task. It takes the question and the core content of the documents as input and outputs the optimal structure type. The router considers five candidate structure types: table, graph, algorithm, catalogue, and chunk, each suited for different kinds of knowledge-intensive tasks.</p></li></ol><p>The selection of the optimal structure type is crucial, as it directly impacts the effectiveness of the subsequent modules. To train the router, the authors propose a novel method based on the Decision Transformer with Preference Optimization (DPO) algorithm, which follows reinforcement learning principles without requiring additional reward models. The training data for the router is generated through a synthesizing-simulating-judging pipeline, which creates high-quality synthetic preference pairs for various tasks and structure types.</p><ol start="2"><li><p>Scattered Knowledge Structurizer: Once the optimal structure type is identified, the scattered knowledge structurizer comes into play. This module is responsible for extracting relevant information scattered across the raw documents and reconstructing it into structured knowledge in the chosen format. The structurizer leverages the powerful understanding and generation capabilities of LLMs to perform this complex task.</p></li></ol><p>The structurizer takes the question, the selected structure type, and each raw document as input. It then extracts the structured knowledge from the document and generates a description of the structured knowledge. The output structured knowledge is collected and combined to form the overall structured knowledge for the given task.</p><ol start="3"><li><p>Structured Knowledge Utilizer: The final module in the StructRAG framework is the structured knowledge utilizer, which performs reasoning to answer the question based on the constructed structured knowledge. This module is designed to handle complex, combinatorial questions that may hinder the direct identification and utilization of relevant information.</p></li></ol><p>The utilizer employs an LLM-based approach to facilitate question decomposition, precise knowledge extraction, and final answer inference. It first breaks down the original question into several simpler sub-questions based on the overall description of the structured knowledge. Then, it extracts precise knowledge for each sub-question from the structured knowledge. Finally, the utilizer integrates all the sub-questions and their corresponding precise knowledge to generate the final answer.</p><h2>Training the Hybrid Structure Router</h2><p>The performance of the hybrid structure router is critical to the overall effectiveness of the StructRAG framework. To train the router, the authors propose a novel method that combines a synthesizing-simulating-judging pipeline for generating training data and the DPO algorithm for preference training.</p>
      <p>
          <a href="https://www.llmwatch.com/p/structrag-succeeding-where-graph">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[SFR-RAG: How Open AI Can Beat OpenAI]]></title><description><![CDATA[Advancing Contextual Understanding in Large Language Models]]></description><link>https://www.llmwatch.com/p/sfr-rag-how-open-ai-can-beat-openai</link><guid isPermaLink="false">https://www.llmwatch.com/p/sfr-rag-how-open-ai-can-beat-openai</guid><dc:creator><![CDATA[Pascal Biese]]></dc:creator><pubDate>Wed, 25 Sep 2024 15:33:23 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/4247e98a-561a-4ee5-8a86-b5e4b863248f_1037x778.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Retrieval Augmented Generation (RAG) has emerged as a critical paradigm for enhancing the capabilities of large language models (LLMs). The recently introduced SFR-RAG model, developed by researchers at Salesforce AI Research, represents a promising direction we&#8217;ve seen evolving this year: small(er) models closing the performance gap between proprietary and open AI. </p><p>This article will explore the key features of SFR-RAG, its performance on various benchmarks, and a step by step explanation of how it improves RAG performance.</p><h2>Key Features of SFR-RAG</h2><p><a href="https://arxiv.org/abs/2409.09916">SFR-RAG</a> is a 9-billion parameter language model specifically designed to excel in RAG applications. The model's primary goal is to faithfully and comprehensively understand provided context and user questions, avoid hallucination, handle challenging scenarios, perform complex reasoning, and produce reliable citations. Let's break down the key aspects of SFR-RAG and how it achieves these objectives.</p><h3>Novel Chat Template</h3><p>Traditional LLMs typically use three roles in their conversational structure: System, User, and Assistant. SFR-RAG expands on this by adding two new roles: Thought and Observation. </p><p>This comes with the following benefits:</p><p>a) Role Clarification: By introducing separate roles for Thought and Observation, SFR-RAG creates a clearer structure for different types of information. This helps the model distinguish between its internal reasoning process (Thought) and external information (Observation).</p><p>b) Easier Masking During Training: The new template allows for more precise control over which parts of the conversation should be included in the training loss. Specifically, System, User, and Observation turns can be masked out, while Thought and Assistant turns are included in the fine-tuning process.</p><p>c) Enhanced Security: The separation of roles facilitates better instruction hierarchy enforcement. This makes the model more resistant to potential jailbreaks or malicious instructions injected through User or Observation turns.</p><p>d) Improved Developer Control: The new template streamlines the process of building reliable and secure RAG applications. Developers can more easily control which parts of the internal processing to display or hide from end-users.</p><p>e) Consistent Function Calling: By designating a specific role (Thought) for internal reasoning and tool use syntax, SFR-RAG avoids the need to parse custom keywords from the Assistant output, leading to more reliable function calling.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Zr8J!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8666fe63-01a4-449b-bc27-f351b88a63ab_1053x601.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Zr8J!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8666fe63-01a4-449b-bc27-f351b88a63ab_1053x601.png 424w, https://substackcdn.com/image/fetch/$s_!Zr8J!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8666fe63-01a4-449b-bc27-f351b88a63ab_1053x601.png 848w, https://substackcdn.com/image/fetch/$s_!Zr8J!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8666fe63-01a4-449b-bc27-f351b88a63ab_1053x601.png 1272w, https://substackcdn.com/image/fetch/$s_!Zr8J!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8666fe63-01a4-449b-bc27-f351b88a63ab_1053x601.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Zr8J!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8666fe63-01a4-449b-bc27-f351b88a63ab_1053x601.png" width="1053" height="601" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8666fe63-01a4-449b-bc27-f351b88a63ab_1053x601.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:601,&quot;width&quot;:1053,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:172609,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Zr8J!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8666fe63-01a4-449b-bc27-f351b88a63ab_1053x601.png 424w, https://substackcdn.com/image/fetch/$s_!Zr8J!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8666fe63-01a4-449b-bc27-f351b88a63ab_1053x601.png 848w, https://substackcdn.com/image/fetch/$s_!Zr8J!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8666fe63-01a4-449b-bc27-f351b88a63ab_1053x601.png 1272w, https://substackcdn.com/image/fetch/$s_!Zr8J!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8666fe63-01a4-449b-bc27-f351b88a63ab_1053x601.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Example of the chat format used by SFR-RAG. <a href="https://arxiv.org/abs/2409.09916">Source</a>.</figcaption></figure></div><h3>Comprehensive Fine-tuning Process</h3><p>The model underwent an extensive fine-tuning process designed to enhance its contextual understanding and generation abilities. This process focused on several key capabilities:</p><ol><li><p>Extracting Relevant Information: SFR-RAG is trained to efficiently extract pertinent information from long contexts. This is crucial for RAG applications where the model needs to sift through large amounts of retrieved data.</p></li><li><p>Recognizing Information Gaps: The model is trained to identify when relevant information is lacking in the provided context. This helps prevent hallucination by encouraging the model to abstain from generating responses when it lacks sufficient information.</p></li><li><p>Handling Conflicting Information: SFR-RAG is equipped to recognize and deal with potentially conflicting information in contextual passages. This is essential for real-world applications where retrieved information may be inconsistent or contradictory.</p></li><li><p>Resilience to Distractions: The fine-tuning process includes exposure to distracting, counter-intuitive, or out-of-distribution content. This helps the model maintain focus on relevant information even in the presence of noise.</p></li><li><p>Diverse Instruction Following: By using extensive instruction-following data that mimics real-world retrieval question answering applications, SFR-RAG is trained to handle a wide variety of tasks and query types.</p></li></ol><h2>Model Performance</h2><p>To evaluate SFR-RAG's performance, the researchers introduced <a href="https://github.com/SalesforceAIResearch/SFR-RAG/blob/main/README_ContextualBench.md">ContextualBench</a>, a comprehensive evaluation suite comprising seven popular contextual question-answering tasks. This standardized benchmark allows for consistent comparison across different models and studies.</p>
      <p>
          <a href="https://www.llmwatch.com/p/sfr-rag-how-open-ai-can-beat-openai">
              Read more
          </a>
      </p>
   ]]></content:encoded></item></channel></rss>