<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[Rohan's Bytes: AI Paper Explained]]></title><description><![CDATA[Here is the collection of all my explanations of recent top AI papers.]]></description><link>https://www.rohan-paul.com/s/ai-paper-podcasts</link><image><url>https://substackcdn.com/image/fetch/$s_!q7Ea!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F20030c9b-4180-453e-9ef3-f42abd8f9de5_1200x1200.png</url><title>Rohan&apos;s Bytes: AI Paper Explained</title><link>https://www.rohan-paul.com/s/ai-paper-podcasts</link></image><generator>Substack</generator><lastBuildDate>Fri, 01 May 2026 16:31:10 GMT</lastBuildDate><atom:link href="https://www.rohan-paul.com/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Rohan Paul]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[rohanpaul@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[rohanpaul@substack.com]]></itunes:email><itunes:name><![CDATA[Rohan Paul]]></itunes:name></itunes:owner><itunes:author><![CDATA[Rohan Paul]]></itunes:author><googleplay:owner><![CDATA[rohanpaul@substack.com]]></googleplay:owner><googleplay:email><![CDATA[rohanpaul@substack.com]]></googleplay:email><googleplay:author><![CDATA[Rohan Paul]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[Mastering Long-Form Reasoning with DAPO]]></title><description><![CDATA[The paper proposes, the Decoupled Clip and Dynamic sAmpling Policy Optimization (DAPO) algorithm, and fully open-source a state-of-the-art large-scale RL system]]></description><link>https://www.rohan-paul.com/p/mastering-long-form-reasoning-with</link><guid isPermaLink="false">https://www.rohan-paul.com/p/mastering-long-form-reasoning-with</guid><dc:creator><![CDATA[Rohan Paul]]></dc:creator><pubDate>Fri, 28 Mar 2025 23:41:16 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!tOmQ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa877ca71-a1ff-49ca-b4e9-78869cb731e6_898x390.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!tOmQ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa877ca71-a1ff-49ca-b4e9-78869cb731e6_898x390.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!tOmQ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa877ca71-a1ff-49ca-b4e9-78869cb731e6_898x390.png 424w, https://substackcdn.com/image/fetch/$s_!tOmQ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa877ca71-a1ff-49ca-b4e9-78869cb731e6_898x390.png 848w, https://substackcdn.com/image/fetch/$s_!tOmQ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa877ca71-a1ff-49ca-b4e9-78869cb731e6_898x390.png 1272w, https://substackcdn.com/image/fetch/$s_!tOmQ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa877ca71-a1ff-49ca-b4e9-78869cb731e6_898x390.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!tOmQ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa877ca71-a1ff-49ca-b4e9-78869cb731e6_898x390.png" width="898" height="390" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/a877ca71-a1ff-49ca-b4e9-78869cb731e6_898x390.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:390,&quot;width&quot;:898,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:75704,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.rohan-paul.com/i/160102641?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa877ca71-a1ff-49ca-b4e9-78869cb731e6_898x390.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!tOmQ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa877ca71-a1ff-49ca-b4e9-78869cb731e6_898x390.png 424w, https://substackcdn.com/image/fetch/$s_!tOmQ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa877ca71-a1ff-49ca-b4e9-78869cb731e6_898x390.png 848w, https://substackcdn.com/image/fetch/$s_!tOmQ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa877ca71-a1ff-49ca-b4e9-78869cb731e6_898x390.png 1272w, https://substackcdn.com/image/fetch/$s_!tOmQ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa877ca71-a1ff-49ca-b4e9-78869cb731e6_898x390.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The <strong><a href="https://arxiv.org/pdf/2503.14476">paper proposes</a></strong>, the Decoupled Clip and Dynamic sAmpling Policy Optimization (DAPO) algorithm, and fully open-source a state-of-the-art large-scale RL system</p><h4>Problem and Motivation</h4><p>Large Language Models often fail to maintain stable performance when fine-tuned with Reinforcement Learning methods for long chain-of-thought reasoning. DeepSeek-R1 introduced large-scale RL to boost reasoning but kept important details private, making replication difficult. DAPO tackles the same challenge with full openness. It closes the performance gap by introducing <strong>Decoupled Clip and Dynamic sAmpling Policy Optimization (DAPO)</strong>. It achieves 50 points on the AIME 2024 math benchmark using the Qwen2.5-32B base model and outperforms the previous 32B DeepSeek-R1 results while requiring fewer training steps.</p><h4>Core Algorithm</h4><p>DAPO builds on GRPO (Group Relative Policy Optimization) but removes a KL divergence term that was unneeded. It also uses a <strong>rule-based reward</strong>: if the answer is correct, reward is 1; otherwise, reward is -1. The advantage function compares each sampled response against others. The policy is updated in smaller, safer steps to stabilize training.</p><h4>Background on GRPO</h4><p><strong>GRPO (Group Relative Policy Optimization)</strong> is a reinforcement learning method that handles multiple responses (or &#8220;action samples&#8221;) per prompt. It calculates the <strong>advantage</strong> of each response relative to other responses sampled under the same prompt. The model then updates its parameters based on these advantage values, ensuring that high-reward responses become more probable and low-reward ones become less probable.</p><h4>The KL Divergence Term</h4><p>In some RL methods (e.g., PPO), a <strong>KL divergence</strong> term is added during optimization. The goal is to keep the updated model from drifting too far from a reference model. This prevents the policy from &#8220;over-updating&#8221; and becoming unstable.</p><h4>Why DAPO Removes KL</h4><p><strong>DAPO</strong> inherits most of GRPO&#8217;s mechanics&#8212;sampling multiple responses, computing advantages, clipping updates&#8212;but <strong>removes the KL component</strong>. This is because DAPO targets <strong>long chain-of-thought reasoning</strong>, where the model&#8217;s output distribution often shifts significantly from the original (pretrained) distribution. Constraining it with KL divergence would hamper the model from exploring these bigger changes. DAPO lets the model adapt more freely without being forced to stay close to the original behavior.</p><h4>Impact</h4><ol><li><p><strong>Fewer Constraints on Exploration</strong>: Without KL, DAPO does not push the model back to a &#8220;safe zone&#8221; near the pretrained distribution.</p></li><li><p><strong>Better for Long-Form Outputs</strong>: Long chain-of-thought solutions require tokens that might be unlikely under the original model. KL could suppress them and reduce reasoning improvements.</p></li><li><p><strong>Maintains Stability Elsewhere</strong>: DAPO still uses clipping (through techniques like Clip-Higher) to keep updates controlled. This keeps training stable without a KL penalty.</p></li></ol><p>Overall, <strong>DAPO builds on GRPO</strong> but <strong>drops the KL divergence</strong> to allow deeper updates in chain-of-thought RL, avoiding the reference-model tether and enabling more extensive exploration.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!giYs!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faaec7a20-dfe1-430a-8a8d-674233be8d2e_896x478.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!giYs!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faaec7a20-dfe1-430a-8a8d-674233be8d2e_896x478.png 424w, https://substackcdn.com/image/fetch/$s_!giYs!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faaec7a20-dfe1-430a-8a8d-674233be8d2e_896x478.png 848w, https://substackcdn.com/image/fetch/$s_!giYs!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faaec7a20-dfe1-430a-8a8d-674233be8d2e_896x478.png 1272w, https://substackcdn.com/image/fetch/$s_!giYs!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faaec7a20-dfe1-430a-8a8d-674233be8d2e_896x478.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!giYs!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faaec7a20-dfe1-430a-8a8d-674233be8d2e_896x478.png" width="896" height="478" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/aaec7a20-dfe1-430a-8a8d-674233be8d2e_896x478.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:478,&quot;width&quot;:896,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:118771,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.rohan-paul.com/i/160102641?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faaec7a20-dfe1-430a-8a8d-674233be8d2e_896x478.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!giYs!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faaec7a20-dfe1-430a-8a8d-674233be8d2e_896x478.png 424w, https://substackcdn.com/image/fetch/$s_!giYs!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faaec7a20-dfe1-430a-8a8d-674233be8d2e_896x478.png 848w, https://substackcdn.com/image/fetch/$s_!giYs!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faaec7a20-dfe1-430a-8a8d-674233be8d2e_896x478.png 1272w, https://substackcdn.com/image/fetch/$s_!giYs!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faaec7a20-dfe1-430a-8a8d-674233be8d2e_896x478.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h4>Four Key Techniques of DAPO</h4><ul><li><p><strong>Clip-Higher</strong></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!EePC!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79b73be7-3317-4952-9c86-b9ed98422e82_871x253.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!EePC!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79b73be7-3317-4952-9c86-b9ed98422e82_871x253.png 424w, https://substackcdn.com/image/fetch/$s_!EePC!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79b73be7-3317-4952-9c86-b9ed98422e82_871x253.png 848w, https://substackcdn.com/image/fetch/$s_!EePC!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79b73be7-3317-4952-9c86-b9ed98422e82_871x253.png 1272w, https://substackcdn.com/image/fetch/$s_!EePC!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79b73be7-3317-4952-9c86-b9ed98422e82_871x253.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!EePC!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79b73be7-3317-4952-9c86-b9ed98422e82_871x253.png" width="871" height="253" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/79b73be7-3317-4952-9c86-b9ed98422e82_871x253.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:253,&quot;width&quot;:871,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:74896,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.rohan-paul.com/i/160102641?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79b73be7-3317-4952-9c86-b9ed98422e82_871x253.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!EePC!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79b73be7-3317-4952-9c86-b9ed98422e82_871x253.png 424w, https://substackcdn.com/image/fetch/$s_!EePC!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79b73be7-3317-4952-9c86-b9ed98422e82_871x253.png 848w, https://substackcdn.com/image/fetch/$s_!EePC!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79b73be7-3317-4952-9c86-b9ed98422e82_871x253.png 1272w, https://substackcdn.com/image/fetch/$s_!EePC!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79b73be7-3317-4952-9c86-b9ed98422e82_871x253.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Traditional RL clipping sets a single clipping range around 1, for example [0.8, 1.2], to keep the new probability (p_new) close to the old probability (p_old). This prevents overly large updates.</p><p><strong>Why standard clipping is restrictive</strong><br>When the clip range is small (say 0.2), a low-probability token with p_old = 0.01 can only rise to p_new = 0.012. That may be too small. This holds back tokens that might become important but started with low probability. Those tokens stay suppressed, limiting the model&#8217;s exploration. Over time, the model overconfidently focuses on a narrow set of responses, called <strong>entropy collapse</strong>.</p><p><strong>What Clip-Higher changes</strong><br>Clip-Higher separates the lower bound and upper bound. The lower bound remains near 1 - epsilon_low (for example 0.8), but the upper bound is set higher (for example 1 + epsilon_high = 1.28). This higher ceiling helps low-probability tokens climb further within a single training step.</p><p><strong>Why it matters</strong></p><ul><li><p><strong>More exploration</strong>: Tokens once deemed improbable can now gain enough probability mass to appear more often.</p></li><li><p><strong>Less collapse</strong>: The model avoids locking onto a few tokens. Entropy remains healthier, so the system explores more reasoning paths.</p></li><li><p><strong>Better diversity</strong>: A richer token distribution emerges, leading to better coverage of potential solutions.</p></li></ul><p>By decoupling the clip bounds, Clip-Higher prevents the model from stagnating on a narrow set of responses, allowing it to discover better reasoning paths and maintain diversity.</p></li><li><p><strong>Dynamic Sampling</strong></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!wBzj!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa0fafc2-dbe9-4a8a-9fa7-0379449aae68_877x437.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!wBzj!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa0fafc2-dbe9-4a8a-9fa7-0379449aae68_877x437.png 424w, https://substackcdn.com/image/fetch/$s_!wBzj!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa0fafc2-dbe9-4a8a-9fa7-0379449aae68_877x437.png 848w, https://substackcdn.com/image/fetch/$s_!wBzj!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa0fafc2-dbe9-4a8a-9fa7-0379449aae68_877x437.png 1272w, https://substackcdn.com/image/fetch/$s_!wBzj!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa0fafc2-dbe9-4a8a-9fa7-0379449aae68_877x437.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!wBzj!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa0fafc2-dbe9-4a8a-9fa7-0379449aae68_877x437.png" width="877" height="437" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/aa0fafc2-dbe9-4a8a-9fa7-0379449aae68_877x437.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:437,&quot;width&quot;:877,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:103320,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.rohan-paul.com/i/160102641?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa0fafc2-dbe9-4a8a-9fa7-0379449aae68_877x437.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!wBzj!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa0fafc2-dbe9-4a8a-9fa7-0379449aae68_877x437.png 424w, https://substackcdn.com/image/fetch/$s_!wBzj!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa0fafc2-dbe9-4a8a-9fa7-0379449aae68_877x437.png 848w, https://substackcdn.com/image/fetch/$s_!wBzj!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa0fafc2-dbe9-4a8a-9fa7-0379449aae68_877x437.png 1272w, https://substackcdn.com/image/fetch/$s_!wBzj!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Faa0fafc2-dbe9-4a8a-9fa7-0379449aae68_877x437.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><strong><br></strong>When every sampled response for a prompt is correct, the advantage for that prompt becomes zero. This stalls training for those samples. DAPO filters out these zero-signal samples by swapping them with new prompts until each batch contains prompts whose responses yield nonzero advantage. This maintains strong training signals without slowing throughput.</p></li><li><p><strong>Token-Level Policy Gradient Loss</strong></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Fepr!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8db10d87-10f4-4d45-bd90-47bba15e8748_887x460.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Fepr!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8db10d87-10f4-4d45-bd90-47bba15e8748_887x460.png 424w, https://substackcdn.com/image/fetch/$s_!Fepr!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8db10d87-10f4-4d45-bd90-47bba15e8748_887x460.png 848w, https://substackcdn.com/image/fetch/$s_!Fepr!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8db10d87-10f4-4d45-bd90-47bba15e8748_887x460.png 1272w, https://substackcdn.com/image/fetch/$s_!Fepr!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8db10d87-10f4-4d45-bd90-47bba15e8748_887x460.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Fepr!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8db10d87-10f4-4d45-bd90-47bba15e8748_887x460.png" width="887" height="460" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8db10d87-10f4-4d45-bd90-47bba15e8748_887x460.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:460,&quot;width&quot;:887,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:141463,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.rohan-paul.com/i/160102641?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8db10d87-10f4-4d45-bd90-47bba15e8748_887x460.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Fepr!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8db10d87-10f4-4d45-bd90-47bba15e8748_887x460.png 424w, https://substackcdn.com/image/fetch/$s_!Fepr!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8db10d87-10f4-4d45-bd90-47bba15e8748_887x460.png 848w, https://substackcdn.com/image/fetch/$s_!Fepr!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8db10d87-10f4-4d45-bd90-47bba15e8748_887x460.png 1272w, https://substackcdn.com/image/fetch/$s_!Fepr!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8db10d87-10f4-4d45-bd90-47bba15e8748_887x460.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>11GRPO averages losses at the sample level. Longer responses get diluted weight, so their tokens are under-penalized or under-rewarded. DAPO aggregates over all tokens across the entire batch. This ensures each token&#8217;s contribution scales with its presence. Long high-quality reasoning gets reinforced more accurately, and long poor-quality outputs are penalized properly.</p></li><li><p><strong>Overlong Reward Shaping</strong></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!fU-b!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ad5434b-18fa-4ca6-8cb8-4f116dfe48ae_839x409.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!fU-b!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ad5434b-18fa-4ca6-8cb8-4f116dfe48ae_839x409.png 424w, https://substackcdn.com/image/fetch/$s_!fU-b!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ad5434b-18fa-4ca6-8cb8-4f116dfe48ae_839x409.png 848w, https://substackcdn.com/image/fetch/$s_!fU-b!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ad5434b-18fa-4ca6-8cb8-4f116dfe48ae_839x409.png 1272w, https://substackcdn.com/image/fetch/$s_!fU-b!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ad5434b-18fa-4ca6-8cb8-4f116dfe48ae_839x409.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!fU-b!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ad5434b-18fa-4ca6-8cb8-4f116dfe48ae_839x409.png" width="839" height="409" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/1ad5434b-18fa-4ca6-8cb8-4f116dfe48ae_839x409.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:409,&quot;width&quot;:839,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:53723,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.rohan-paul.com/i/160102641?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ad5434b-18fa-4ca6-8cb8-4f116dfe48ae_839x409.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!fU-b!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ad5434b-18fa-4ca6-8cb8-4f116dfe48ae_839x409.png 424w, https://substackcdn.com/image/fetch/$s_!fU-b!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ad5434b-18fa-4ca6-8cb8-4f116dfe48ae_839x409.png 848w, https://substackcdn.com/image/fetch/$s_!fU-b!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ad5434b-18fa-4ca6-8cb8-4f116dfe48ae_839x409.png 1272w, https://substackcdn.com/image/fetch/$s_!fU-b!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ad5434b-18fa-4ca6-8cb8-4f116dfe48ae_839x409.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>In RL, responses sometimes exceed a maximum token count set by the training system. The system chops off any excess tokens, so the response is &#8220;truncated.&#8221;</p><p>When a response is truncated, some training setups automatically assign it a &#8220;bad&#8221; reward. That negative signal may be misleading. The reasoning process might be correct but just too verbose. Penalizing it with a flat negative reward confuses the model.</p><p><strong>Overlong Filtering</strong></p><ul><li><p>DAPO skips using these truncated samples when updating the model.</p></li><li><p>The model does not see an incorrect penalty for what could be a valid reasoning chain.</p></li><li><p>This avoids mixing &#8220;overly long&#8221; with &#8220;low-quality.&#8221;</p></li></ul><p><strong>Soft Overlong Punishment</strong></p><ul><li><p>DAPO applies a small penalty once the response length goes past a certain threshold, and the penalty grows if the output gets even longer.</p></li><li><p>The aim is to nudge the model to be concise without giving it a harsh, all-or-nothing punishment.</p></li><li><p>The model still keeps any correct logic it used so far, but sees an incentive to shorten future outputs.</p></li></ul><p>This two-step method keeps training signals clear. Good reasoning is reinforced, while the model is gently encouraged to avoid excess length.</p></li></ul><h2>The Algorithm of DAPO</h2><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Kx1v!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77efe9e1-b4b4-41c1-8c81-6b9add80661c_852x342.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Kx1v!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77efe9e1-b4b4-41c1-8c81-6b9add80661c_852x342.png 424w, https://substackcdn.com/image/fetch/$s_!Kx1v!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77efe9e1-b4b4-41c1-8c81-6b9add80661c_852x342.png 848w, https://substackcdn.com/image/fetch/$s_!Kx1v!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77efe9e1-b4b4-41c1-8c81-6b9add80661c_852x342.png 1272w, https://substackcdn.com/image/fetch/$s_!Kx1v!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77efe9e1-b4b4-41c1-8c81-6b9add80661c_852x342.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Kx1v!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77efe9e1-b4b4-41c1-8c81-6b9add80661c_852x342.png" width="852" height="342" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/77efe9e1-b4b4-41c1-8c81-6b9add80661c_852x342.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:342,&quot;width&quot;:852,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:91007,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.rohan-paul.com/i/160102641?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77efe9e1-b4b4-41c1-8c81-6b9add80661c_852x342.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Kx1v!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77efe9e1-b4b4-41c1-8c81-6b9add80661c_852x342.png 424w, https://substackcdn.com/image/fetch/$s_!Kx1v!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77efe9e1-b4b4-41c1-8c81-6b9add80661c_852x342.png 848w, https://substackcdn.com/image/fetch/$s_!Kx1v!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77efe9e1-b4b4-41c1-8c81-6b9add80661c_852x342.png 1272w, https://substackcdn.com/image/fetch/$s_!Kx1v!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77efe9e1-b4b4-41c1-8c81-6b9add80661c_852x342.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><strong>Input</strong></p><ul><li><p><strong>&#960;&#952;</strong>: The policy model we are currently training.</p></li><li><p><strong>R</strong>: The reward function that assigns a score to each generated response.</p></li><li><p><strong>D</strong>: The dataset (or collection of prompts) we use to query the model.</p></li><li><p><strong>&#949;_low, &#949;_high</strong>: The two key hyperparameters that set the lower and upper clip range.</p></li></ul><p><strong>Main Loop (step = 1 to M)</strong></p><ol><li><p><strong>Sample a batch Db from D</strong></p><ul><li><p>We pick a batch of prompts from our training set D. These are the questions (or tasks) for which the model will generate responses and receive rewards.</p></li></ul></li><li><p><strong>Update the old policy model &#960;&#952;_old &#8592; &#960;&#952;</strong></p><ul><li><p>We make a copy of our current policy parameters so we can compare old probabilities to new ones during policy updates.</p></li></ul></li><li><p><strong>Sample G outputs {oi} for each question in Db using &#960;&#952;_old</strong></p><ul><li><p>For every prompt q in this batch, we ask the old policy model to generate G different responses.</p></li><li><p>These multiple samples let us measure relative quality among responses for the same prompt.</p></li></ul></li><li><p><strong>Compute rewards {ri} for each output by running R</strong></p><ul><li><p>Each response gets a reward, for example 1 if the answer is correct, -1 if incorrect.</p></li><li><p>Other forms of reward can also be applied, as long as they map each response to a numeric score.</p></li></ul></li><li><p><strong>Filter out oi and add the remaining to the dynamic sampling buffer</strong></p><ul><li><p>Dynamic Sampling means we only keep prompts that do not yield all-correct or all-wrong outputs.</p></li><li><p>If all G responses for a prompt are correct, there is no training signal (the advantages become zero). If all are wrong, we often skip them for a similar reason.</p></li><li><p>We keep only prompts that have meaningful variation in correctness and store them in a buffer.</p></li></ul></li><li><p><strong>Check buffer size nb</strong></p><ul><li><p>If the buffer has fewer than N items (N is a chosen threshold), we skip the update for now.</p></li><li><p>This ensures each update sees enough training signals.</p></li></ul></li><li><p><strong>Compute advantages A_i,t for each token</strong></p><ul><li><p>For each response that survived filtering, we compare the reward of that response to the reward of other responses for the same prompt.</p></li><li><p>This advantage is token-level, meaning we assign each token&#8217;s share of credit for the overall reward.</p></li></ul></li><li><p><strong>Perform multiple gradient updates</strong></p><ul><li><p>We loop over a small number of gradient-update steps (&#956;).</p></li><li><p>In each update, we <strong>maximize the DAPO objective</strong>, which applies clipping (with &#949;_low and &#949;_high) at each token.</p></li><li><p>This step nudges the policy toward higher-reward responses and away from lower-reward responses.</p></li><li><p>The key difference vs. older algorithms is the <strong>decoupled</strong> clip range and the <strong>token-level</strong> averaging, plus dynamic sampling.</p></li></ul></li></ol><p><strong>Output</strong></p><ul><li><p>The final trained policy &#960;&#952; after M iterations.</p></li><li><p>It has learned to produce more correct or higher-reward outputs by repeatedly sampling, comparing, and updating based on these feedback signals.</p></li></ul><h4>Dataset and Training</h4><p>The team curated <strong>DAPO-Math-17K</strong>, a math dataset with integer-answer problems to simplify rule-based scoring. Qwen2.5-32B was used as the base model. The RL training starts from that checkpoint, applies DAPO, and samples multiple responses per prompt. Each step updates the model by maximizing the difference in reward between good and bad responses.</p><h4>Results</h4><p>Accuracy on the AIME 2024 benchmark climbed from 30 points under naive GRPO to 50 points under DAPO. Dynamic Sampling, Clip-Higher, Overlong Reward Shaping, and Token-Level Loss each contributed to closing the gap. DAPO needed roughly half the number of training steps compared to the previous 32B DeepSeek-R1.</p><h4>Observations on Reasoning Behaviors</h4><p>During training, the model began to <strong>reflect</strong> on its reasoning and revise steps mid-solution, a behavior absent initially. This emergence suggests DAPO fosters exploration of more complex reasoning paths.</p><p>DAPO&#8217;s <strong><a href="https://github.com/BytedTsinghua-SIA/DAPO">codebase and dataset</a></strong> are fully open-sourced, enabling reproducibility of large-scale RL training for advanced reasoning tasks. These techniques stabilize training, preserve exploration, and encourage correct long-form reasoning in LLMs.</p>]]></content:encoded></item><item><title><![CDATA[Particle-based inference-time scaling helps smaller LLMs reach or exceed top-tier closed-source models on math benchmarks]]></title><description><![CDATA[Paper - https://arxiv.org/abs/2502.01618]]></description><link>https://www.rohan-paul.com/p/particle-based-inference-time-scaling</link><guid isPermaLink="false">https://www.rohan-paul.com/p/particle-based-inference-time-scaling</guid><dc:creator><![CDATA[Rohan Paul]]></dc:creator><pubDate>Mon, 24 Mar 2025 15:28:27 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!O4vD!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F75547d06-2056-45b1-82aa-572ac96837f0_872x868.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!O4vD!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F75547d06-2056-45b1-82aa-572ac96837f0_872x868.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!O4vD!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F75547d06-2056-45b1-82aa-572ac96837f0_872x868.png 424w, https://substackcdn.com/image/fetch/$s_!O4vD!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F75547d06-2056-45b1-82aa-572ac96837f0_872x868.png 848w, https://substackcdn.com/image/fetch/$s_!O4vD!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F75547d06-2056-45b1-82aa-572ac96837f0_872x868.png 1272w, https://substackcdn.com/image/fetch/$s_!O4vD!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F75547d06-2056-45b1-82aa-572ac96837f0_872x868.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!O4vD!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F75547d06-2056-45b1-82aa-572ac96837f0_872x868.png" width="872" height="868" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/75547d06-2056-45b1-82aa-572ac96837f0_872x868.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:868,&quot;width&quot;:872,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:329664,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://rohanpaul.substack.com/i/159754812?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F75547d06-2056-45b1-82aa-572ac96837f0_872x868.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!O4vD!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F75547d06-2056-45b1-82aa-572ac96837f0_872x868.png 424w, https://substackcdn.com/image/fetch/$s_!O4vD!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F75547d06-2056-45b1-82aa-572ac96837f0_872x868.png 848w, https://substackcdn.com/image/fetch/$s_!O4vD!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F75547d06-2056-45b1-82aa-572ac96837f0_872x868.png 1272w, https://substackcdn.com/image/fetch/$s_!O4vD!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F75547d06-2056-45b1-82aa-572ac96837f0_872x868.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Paper - <a href="https://arxiv.org/abs/2502.01618">https://arxiv.org/abs/2502.01618</a></figcaption></figure></div><p><strong>Paper - https://arxiv.org/abs/2502.01618</strong></p><p>They want to improve small LLMs&#8217; reasoning accuracy by allocating more compute at inference time.</p><p>Propose treating inference-time scaling as probabilistic inference over a <strong>state-space model (SSM)</strong>, using <strong>particle-based sampling</strong> to handle imperfect reward models.</p><h4>&#127991;&#65039; Overview of the Method</h4><p>They define a <strong>transition</strong> via the policy model and treat acceptance as a <strong>likelihood</strong> from a <strong>process reward model (PRM)</strong>. They then apply <strong>particle filtering (PF)</strong> to sample from the posterior distribution instead of directly optimizing the reward, reducing reward hacking and preserving exploration. Each partial solution&#8217;s reward guides resampling, promoting higher-scoring paths while keeping diversity.</p><h4>&#127991;&#65039; Particle Filtering</h4><p>PF starts with several partial answers (particles). Each step:</p><ul><li><p>Draws new tokens from the LLM.</p></li><li><p>Scores them via the PRM.</p></li><li><p>Weights and resamples the particles proportionally to their reward-based scores.</p></li></ul><p>This stochastic approach balances exploitation and exploration, unlike deterministic search methods that can overfit to reward flaws.</p><h4>&#9881;&#65039; Core Idea of Particle Filtering</h4><p>Particle Filtering maintains multiple &#8220;particles&#8221; (possible scenarios) that evolve step by step. Each particle represents one possible hidden state sequence. Every time new data arrives, particles get <strong>weighted</strong> based on how well they match the data, then <strong>resampled</strong> so higher-weight particles survive. This keeps multiple hypotheses alive until the end.</p><h4>&#128296; Main Steps</h4><ol><li><p><strong>Initialization</strong><br>Draw a batch of random guesses (particles). Assign each a weight of 1/N.</p></li><li><p><strong>Prediction</strong><br>Evolve each particle using the process model (e.g., a language model). This predicts the next state.</p></li><li><p><strong>Weight Update</strong><br>Compare each particle&#8217;s prediction to observed data via the likelihood (e.g., reward model). Multiply its weight by that likelihood.</p></li><li><p><strong>Resampling</strong><br>Randomly pick particles (with replacement) according to their weights to form a new batch. Particles with higher weights get replicated; low-weight ones vanish.</p></li><li><p><strong>Iteration</strong><br>Move to the next time step. Repeat prediction, weight update, and resampling. Continue until all steps are processed.</p></li></ol><h4>&#129513; Why It Works</h4><p>It balances two goals:</p><ul><li><p><strong>Exploit</strong> high-likelihood particles.</p></li><li><p><strong>Explore</strong> by keeping some diversity.</p></li></ul><p>It avoids getting stuck in just one best guess, so it&#8217;s more robust to noisy or imperfect models.</p><div><hr></div><h4>&#127991;&#65039; Multi-Iteration and Parallel Extensions</h4><p>They extend PF with:</p><ul><li><p><strong>Particle Gibbs (PG)</strong>, which reruns PF multiple times, keeping a reference trajectory between runs for improved sampling.</p></li><li><p><strong>Parallel Tempering (PT)</strong>, which runs multiple PF chains at different &#8220;temperatures&#8221; and occasionally swaps their states to combine exploration and exploitation more effectively.</p></li></ul><h4>&#127991;&#65039; Experiments and Results</h4><p>They evaluate on <strong>MATH500</strong> (500 math problems) and <strong>AIME2024</strong> (30 competition problems). Key highlights:</p><ul><li><p><strong>PF</strong> outperforms standard methods like <strong>best-of-n (BoN)</strong>, <strong>weighted BoN (WBoN)</strong>, and <strong>dynamic variable-time search (DVTS)</strong>.</p></li><li><p>PF consistently achieves <strong>4&#8211;16x</strong> better scaling efficiency.</p></li><li><p>With only <strong>4</strong> particles, <strong>Qwen2.5-Math-1.5B-Instruct</strong> surpasses <strong>GPT-4o</strong> on MATH500.</p></li><li><p>With <strong>32</strong> particles, <strong>Qwen2.5-Math-7B-Instruct</strong> matches <strong>o1-preview</strong> on MATH500.</p></li><li><p><strong>Llama-3.1-8B-Instruct</strong> with PF exceeds or equals <strong>Llama-3.1-70B-Instruct</strong> performance.</p></li></ul><h4>&#127991;&#65039; Ablations</h4><p>They tested different <strong>PRMs</strong> (e.g. Qwen2.5-Math-PRM-7B) and various reward-aggregation strategies. A single aggregated step-level reward or a final outcome reward can be used. They also show that moderate LLM sampling temperatures (around 0.8) work best.</p><h4>&#127991;&#65039; Key Insights</h4><p>Framing inference-time scaling as a sampling problem over an approximate reward avoids rigid optimization on a possibly flawed reward. Particle-based sampling maintains a stable &#8220;typical set&#8221; exploration, leading to robust gains on challenging reasoning tasks.</p>]]></content:encoded></item><item><title><![CDATA["Speak Easy: Eliciting Harmful Jailbreaks from LLMs with Simple Interactions"]]></title><description><![CDATA[Below podcast on this paper is generated with Google's Illuminate.]]></description><link>https://www.rohan-paul.com/p/speak-easy-eliciting-harmful-jailbreaks</link><guid isPermaLink="false">https://www.rohan-paul.com/p/speak-easy-eliciting-harmful-jailbreaks</guid><dc:creator><![CDATA[Rohan Paul]]></dc:creator><pubDate>Sun, 16 Feb 2025 14:53:44 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!I0bX!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5a3949a4-4f11-46e4-b931-7180fbc8cefb_1135x871.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!I0bX!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5a3949a4-4f11-46e4-b931-7180fbc8cefb_1135x871.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!I0bX!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5a3949a4-4f11-46e4-b931-7180fbc8cefb_1135x871.png 424w, https://substackcdn.com/image/fetch/$s_!I0bX!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5a3949a4-4f11-46e4-b931-7180fbc8cefb_1135x871.png 848w, https://substackcdn.com/image/fetch/$s_!I0bX!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5a3949a4-4f11-46e4-b931-7180fbc8cefb_1135x871.png 1272w, https://substackcdn.com/image/fetch/$s_!I0bX!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5a3949a4-4f11-46e4-b931-7180fbc8cefb_1135x871.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!I0bX!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5a3949a4-4f11-46e4-b931-7180fbc8cefb_1135x871.png" width="1135" height="871" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/5a3949a4-4f11-46e4-b931-7180fbc8cefb_1135x871.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:871,&quot;width&quot;:1135,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:451716,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!I0bX!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5a3949a4-4f11-46e4-b931-7180fbc8cefb_1135x871.png 424w, https://substackcdn.com/image/fetch/$s_!I0bX!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5a3949a4-4f11-46e4-b931-7180fbc8cefb_1135x871.png 848w, https://substackcdn.com/image/fetch/$s_!I0bX!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5a3949a4-4f11-46e4-b931-7180fbc8cefb_1135x871.png 1272w, https://substackcdn.com/image/fetch/$s_!I0bX!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5a3949a4-4f11-46e4-b931-7180fbc8cefb_1135x871.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><a href="https://arxiv.org/abs/2502.04322">https://arxiv.org/abs/2502.04322</a></p><p>LLMs despite safety alignment, are still vulnerable to jailbreak attacks, eliciting harmful responses. Current jailbreak methods often require technical expertise, but this paper explores if simple, everyday user interactions can also lead to harmful jailbreaks.</p><p>This paper introduces SPEAK EASY, a simple framework simulating realistic user interactions using multi-step and multilingual queries to elicit actionable and informative harmful responses.</p><p>-----</p><p>&#128204; SPEAK EASY cleverly exploits inherent weaknesses in LLMs' (LLMs) safety alignment by mimicking natural user interactions. Multi-step decomposition and multilingual translation effectively bypass surface-level safety filters.</p><p>&#128204; HARMSCORE offers a nuanced evaluation beyond binary Attack Success Rate (ASR). It quantifies the harm potential by focusing on actionability and informativeness, aligning better with human perception of real-world risk.</p><p>&#128204; The framework's modular design is a strength. SPEAK EASY's integration with gradient-based and tree-of-thought methods demonstrates its versatility and practical applicability for robust jailbreak testing.</p><p>----------</p><p>Methods Explored in this Paper &#128295;:</p><p>&#8594; SPEAK EASY framework is proposed to elicit harmful jailbreaks from LLMs using simple interactions.</p><p>&#8594; It simulates real-world user behaviors by employing multi-step query decomposition and multilingual translations.</p><p>&#8594; Given a harmful query, SPEAK EASY first decomposes it into multiple seemingly harmless subqueries.</p><p>&#8594; Each subquery is then translated into multiple languages to exploit multilingual vulnerabilities in LLMs.</p><p>&#8594; Responses from the LLM for translated subqueries are translated back to English.</p><p>&#8594; Actionability and informativeness of each response are scored using fine-tuned response selection models.</p><p>&#8594; The responses with the highest combined actionability and informativeness scores are selected and concatenated to form the final jailbreak response.</p><p>&#8594; HARMSCORE metric is introduced to evaluate jailbreak harmfulness based on actionability and informativeness of the LLM's response, going beyond simple Attack Success Rate (ASR).</p><p>-----</p><p>Key Insights &#128161;:</p><p>&#8594; Actionability and informativeness are identified as key attributes that make a jailbroken LLM response truly harmful and useful for malicious users.</p><p>&#8594; HARMSCORE, a metric measuring actionability and informativeness, aligns better with human judgements of harmfulness than ASR alone, especially for instruction-based harmful queries.</p><p>&#8594; Simple multi-step and multilingual interactions, as simulated by SPEAK EASY, can significantly increase the likelihood of eliciting harmful responses from both proprietary and open-source LLMs.</p><p>&#8594; SPEAK EASY framework is easily integrated with existing jailbreak methods like GCG-T and TAP-T, further enhancing their effectiveness.</p><p>-----</p><p>Results &#128202;:</p><p>&#8594; SPEAK EASY increased the average Attack Success Rate (ASR) of GPT-4o from 0.092 to 0.555 across four benchmarks.</p><p>&#8594; SPEAK EASY increased the average HARMSCORE of GPT-4o from 0.180 to 0.759 across four benchmarks.</p><p>&#8594; Integrating SPEAK EASY with GCG-T and TAP-T significantly improved their ASR and HARMSCORE, with TAP-T+SPEAK EASY achieving over 0.9 ASR on GPT-4o and Llama3.3.</p><p>&#8594; Human evaluation showed HARMSCORE has a Pearson rank correlation of 0.726 with human judgement of harm, which is competitive with GPT-4o based ASR (0.723) and outperforms HarmBench ASR (0.638).</p>]]></content:encoded></item><item><title><![CDATA["SPARC: Subspace-Aware Prompt Adaptation for Robust Continual Learning in LLMs"]]></title><description><![CDATA[Below podcast on this paper is generated with Google's Illuminate.]]></description><link>https://www.rohan-paul.com/p/sparc-subspace-aware-prompt-adaptation</link><guid isPermaLink="false">https://www.rohan-paul.com/p/sparc-subspace-aware-prompt-adaptation</guid><dc:creator><![CDATA[Rohan Paul]]></dc:creator><pubDate>Sun, 16 Feb 2025 14:43:53 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!bmLs!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffa080e41-9e2a-4aa9-ad5d-7b5581c45c87_1090x818.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!bmLs!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffa080e41-9e2a-4aa9-ad5d-7b5581c45c87_1090x818.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!bmLs!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffa080e41-9e2a-4aa9-ad5d-7b5581c45c87_1090x818.png 424w, https://substackcdn.com/image/fetch/$s_!bmLs!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffa080e41-9e2a-4aa9-ad5d-7b5581c45c87_1090x818.png 848w, https://substackcdn.com/image/fetch/$s_!bmLs!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffa080e41-9e2a-4aa9-ad5d-7b5581c45c87_1090x818.png 1272w, https://substackcdn.com/image/fetch/$s_!bmLs!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffa080e41-9e2a-4aa9-ad5d-7b5581c45c87_1090x818.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!bmLs!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffa080e41-9e2a-4aa9-ad5d-7b5581c45c87_1090x818.png" width="1090" height="818" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/fa080e41-9e2a-4aa9-ad5d-7b5581c45c87_1090x818.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:818,&quot;width&quot;:1090,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:397814,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!bmLs!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffa080e41-9e2a-4aa9-ad5d-7b5581c45c87_1090x818.png 424w, https://substackcdn.com/image/fetch/$s_!bmLs!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffa080e41-9e2a-4aa9-ad5d-7b5581c45c87_1090x818.png 848w, https://substackcdn.com/image/fetch/$s_!bmLs!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffa080e41-9e2a-4aa9-ad5d-7b5581c45c87_1090x818.png 1272w, https://substackcdn.com/image/fetch/$s_!bmLs!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffa080e41-9e2a-4aa9-ad5d-7b5581c45c87_1090x818.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><a href="https://arxiv.org/abs/2502.02909">https://arxiv.org/abs/2502.02909</a></p><p>The challenge is enabling LLMs to learn new tasks continually without losing old knowledge, known as catastrophic forgetting, especially under resource constraints. This paper introduces SPARC to tackle this problem efficiently.</p><p>This paper proposes SPARC, a framework using subspace-guided prompt tuning, to allow LLMs to learn continually by adapting prompts in a low-dimensional space derived from PCA.</p><p>-----</p><p>&#128204; SPARC's strength is in efficient continual learning. It cleverly uses PCA to identify task subspaces. Orthogonal prompt initialization then ensures minimal interference. This approach significantly reduces parameter updates, making it highly scalable.</p><p>&#128204; PCA-based subspace identification is key. It allows SPARC to adapt prompts in a low-dimensional space. This focuses learning on task-relevant features. The cosine similarity metric effectively guides prompt reuse or orthogonalization.</p><p>&#128204; By freezing most LLM parameters, SPARC preserves pre-trained knowledge. Fine-tuning only soft prompts achieves strong performance. This highlights the effectiveness of targeted, subspace-aware adaptation for continual learning in LLMs.</p><p>----------</p><p>Methods Explored in this Paper &#128295;:</p><p>&#8594; SPARC uses prompt tuning. It adapts LLMs to new tasks by adding small, trainable vectors called soft prompts to the input. This keeps the main LLM parameters frozen.</p><p>&#8594; Principal Component Analysis, or PCA, is used to find important features of each task's data. PCA reduces the data into a lower-dimensional subspace. This subspace captures the most important information for each task.</p><p>&#8594; SPARC checks if new tasks are similar to old ones using cosine similarity. Cosine similarity measures the overlap between task subspaces. If tasks are similar, SPARC reuses existing prompts. This saves computation and helps transfer knowledge.</p><p>&#8594; For dissimilar tasks, SPARC creates new prompts in orthogonal subspaces. Orthogonal subspaces are independent. This ensures new tasks don't interfere with old ones, preventing forgetting.</p><p>&#8594; SPARC is parameter-efficient. It only trains the soft prompts, which are a tiny fraction of the LLM's parameters. This makes it scalable and resource-friendly. SPARC can also integrate with LoRA for further efficiency.</p><p>-----</p><p>Key Insights &#128161;:</p><p>&#8594; Subspace-guided prompt tuning effectively mitigates catastrophic forgetting in LLMs during continual learning.</p><p>&#8594; Reusing prompts for similar tasks and creating orthogonal prompts for dissimilar tasks balances knowledge retention and adaptation.</p><p>&#8594; PCA helps in identifying task-relevant features and creating efficient, low-dimensional prompts.</p><p>&#8594; SPARC is highly parameter-efficient, requiring fine-tuning of only a small percentage of model parameters. This makes continual learning more practical for large models.</p><p>-----</p><p>Results &#128202;:</p><p>&#8594; SPARC achieves no forgetting in task-incremental learning.</p><p>&#8594; In domain-incremental learning, the average forgetting ratio is below 5%.</p><p>&#8594; SPARC maintains over 97% prior knowledge retention.</p><p>&#8594; SPARC fine-tunes only 0.04% of the model's parameters. Combining SPARC with LoRA, it fine-tunes only 1% of parameters while improving accuracy.</p><p>&#8594; SPARC consistently outperforms baseline fine-tuning methods in knowledge retention and transfer.</p>]]></content:encoded></item><item><title><![CDATA["Show-o Turbo: Towards Accelerated Unified Multimodal Understanding and Generation"]]></title><description><![CDATA[Below podcast on this paper is generated with Google's Illuminate.]]></description><link>https://www.rohan-paul.com/p/show-o-turbo-towards-accelerated</link><guid isPermaLink="false">https://www.rohan-paul.com/p/show-o-turbo-towards-accelerated</guid><dc:creator><![CDATA[Rohan Paul]]></dc:creator><pubDate>Sun, 16 Feb 2025 14:36:32 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!MSdn!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F82f9d35d-7a96-47bb-8b1b-7a85c4ed24bc_975x793.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!MSdn!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F82f9d35d-7a96-47bb-8b1b-7a85c4ed24bc_975x793.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!MSdn!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F82f9d35d-7a96-47bb-8b1b-7a85c4ed24bc_975x793.png 424w, https://substackcdn.com/image/fetch/$s_!MSdn!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F82f9d35d-7a96-47bb-8b1b-7a85c4ed24bc_975x793.png 848w, https://substackcdn.com/image/fetch/$s_!MSdn!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F82f9d35d-7a96-47bb-8b1b-7a85c4ed24bc_975x793.png 1272w, https://substackcdn.com/image/fetch/$s_!MSdn!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F82f9d35d-7a96-47bb-8b1b-7a85c4ed24bc_975x793.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!MSdn!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F82f9d35d-7a96-47bb-8b1b-7a85c4ed24bc_975x793.png" width="975" height="793" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/82f9d35d-7a96-47bb-8b1b-7a85c4ed24bc_975x793.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:793,&quot;width&quot;:975,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:380829,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!MSdn!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F82f9d35d-7a96-47bb-8b1b-7a85c4ed24bc_975x793.png 424w, https://substackcdn.com/image/fetch/$s_!MSdn!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F82f9d35d-7a96-47bb-8b1b-7a85c4ed24bc_975x793.png 848w, https://substackcdn.com/image/fetch/$s_!MSdn!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F82f9d35d-7a96-47bb-8b1b-7a85c4ed24bc_975x793.png 1272w, https://substackcdn.com/image/fetch/$s_!MSdn!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F82f9d35d-7a96-47bb-8b1b-7a85c4ed24bc_975x793.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><a href="https://arxiv.org/abs/2502.05415">https://arxiv.org/abs/2502.05415</a></p><p>The original Show-o model, while versatile for both image and text generation, suffers from slow inference speeds due to its iterative generation processes for both modalities. This paper addresses this inefficiency by introducing Show-o Turbo, a method to accelerate both image and text generation in a unified multimodal model.</p><p>This paper proposes to shorten the generation process in Show-o by applying consistency distillation to its multimodal denoising trajectories. This approach treats both image and text generation as denoising tasks, enabling a unified acceleration strategy.</p><p>-----</p><p>&#128204; Show-o Turbo cleverly unifies text and image generation as denoising. Jacobi decoding enables parallel text token refinement, mirroring image token processing for consistent acceleration.</p><p>&#128204; This paper demonstrates effective generalization of consistency distillation to discrete multimodal generation. Trajectory segmentation stabilizes training and achieves few-step, high-quality multimodal output.</p><p>&#128204; The key innovation is practical acceleration. Show-o Turbo achieves 1.5x speedup in multimodal tasks and maintains strong generation quality with significantly fewer steps.</p><p>----------</p><p>Methods Explored in this Paper &#128295;:</p><p>&#8594; This paper introduces a unified perspective by viewing text generation as a denoising process, similar to image generation in Show-o. This is achieved by applying Jacobi decoding, a parallel text decoding algorithm, to Show-o. Jacobi decoding refines multiple text tokens simultaneously, mimicking the parallel denoising of image tokens.</p><p>&#8594; Show-o Turbo employs consistency distillation to shorten the denoising trajectories for both images and text. Consistency distillation trains Show-o Turbo to map any point on the original Show-o's generation trajectory to the final output in fewer steps. This is done by minimizing the difference between the predictions of the student model (Show-o Turbo) and the teacher model (Show-o) at different points in the generation process.</p><p>&#8594; Trajectory segmentation and curriculum learning are used to improve training. The long denoising trajectory is divided into segments. Consistency distillation is applied within each segment. Training proceeds in stages with decreasing segment lengths, making learning more manageable and improving convergence.</p><p>-----</p><p>Key Insights &#128161;:</p><p>&#8594; Treating text generation as denoising, through parallel decoding methods like Jacobi decoding, allows for a unified acceleration approach for multimodal models. This perspective bridges the gap between image and text generation processes in Show-o.</p><p>&#8594; Consistency distillation, originally developed for continuous diffusion models, can be effectively generalized and extended to accelerate discrete multimodal models like Show-o. This extension enables significant speedups without sacrificing generation quality.</p><p>&#8594; Trajectory segmentation and curriculum learning are crucial for the successful application of consistency distillation to complex models like Show-o. These techniques stabilize training and enhance the effectiveness of distillation by breaking down the learning task into manageable stages.</p><p>-----</p><p>Results &#128202;:</p><p>&#8594; Show-o Turbo achieves a GenEval score of 0.625 in text-to-image generation with only 4 sampling steps, without classifier-free guidance (CFG). This outperforms the original Show-o with 8 steps and CFG, which has a GenEval score of 0.580 (at 5 steps with CFG in Table 1).</p><p>&#8594; In image-to-text generation tasks, Show-o Turbo demonstrates a 1.5x speedup in inference time compared to the original Show-o, while maintaining comparable performance on image description benchmarks like Flickr30K and NoCaps (Table 2).</p><p>&#8594; On multimodal understanding tasks, Show-o Turbo maintains competitive performance on benchmarks like POPE, MME, and MMMU, demonstrating its ability to accelerate without significantly degrading understanding capabilities (Table 2).</p>]]></content:encoded></item><item><title><![CDATA["ScoreFlow: Mastering LLM Agent Workflows via Score-based Preference Optimization"]]></title><description><![CDATA[Below podcast on this paper is generated with Google's Illuminate.]]></description><link>https://www.rohan-paul.com/p/scoreflow-mastering-llm-agent-workflows</link><guid isPermaLink="false">https://www.rohan-paul.com/p/scoreflow-mastering-llm-agent-workflows</guid><dc:creator><![CDATA[Rohan Paul]]></dc:creator><pubDate>Sun, 16 Feb 2025 14:31:02 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!QjI7!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F15c2f09b-b211-4372-ab2b-858f119ccd46_1096x788.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!QjI7!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F15c2f09b-b211-4372-ab2b-858f119ccd46_1096x788.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!QjI7!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F15c2f09b-b211-4372-ab2b-858f119ccd46_1096x788.png 424w, https://substackcdn.com/image/fetch/$s_!QjI7!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F15c2f09b-b211-4372-ab2b-858f119ccd46_1096x788.png 848w, https://substackcdn.com/image/fetch/$s_!QjI7!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F15c2f09b-b211-4372-ab2b-858f119ccd46_1096x788.png 1272w, https://substackcdn.com/image/fetch/$s_!QjI7!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F15c2f09b-b211-4372-ab2b-858f119ccd46_1096x788.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!QjI7!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F15c2f09b-b211-4372-ab2b-858f119ccd46_1096x788.png" width="1096" height="788" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/15c2f09b-b211-4372-ab2b-858f119ccd46_1096x788.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:788,&quot;width&quot;:1096,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:198833,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!QjI7!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F15c2f09b-b211-4372-ab2b-858f119ccd46_1096x788.png 424w, https://substackcdn.com/image/fetch/$s_!QjI7!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F15c2f09b-b211-4372-ab2b-858f119ccd46_1096x788.png 848w, https://substackcdn.com/image/fetch/$s_!QjI7!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F15c2f09b-b211-4372-ab2b-858f119ccd46_1096x788.png 1272w, https://substackcdn.com/image/fetch/$s_!QjI7!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F15c2f09b-b211-4372-ab2b-858f119ccd46_1096x788.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><a href="https://arxiv.org/abs/2502.04306">https://arxiv.org/abs/2502.04306</a></p><p>This paper addresses the challenge of optimizing LLM agent workflows for complex tasks. Existing methods are inflexible, not adaptable and struggle with scalability due to discrete optimization techniques.</p><p>This paper introduces ScoreFlow, a framework using gradient-based optimization in a continuous space. It uses Score-DPO, a new method of direct preference optimization that considers quantitative feedback for improved workflow generation.</p><p>-----</p><p>&#128204; ScoreFlow's strength lies in shifting workflow optimization from discrete search to continuous gradient-based methods. This enables efficient exploration of complex agentic workflows, overcoming limitations of prior discrete approaches like Monte Carlo Tree Search used in AFlow.</p><p>&#128204; Score-DPO innovatively integrates quantitative scores into Direct Preference Optimization. This score-aware preference learning refines workflow generation more effectively than standard DPO, which only uses binary preference pairs, leading to faster convergence and better performance.</p><p>&#128204; The code representation for workflows in ScoreFlow provides a flexible structure. This allows for complex logic, loops, and conditional execution within agent workflows, unlike graph-based methods. This code-based approach enhances adaptability and expressiveness for diverse tasks.</p><p>----------</p><p>Methods Explored in this Paper &#128295;:</p><p>&#8594; ScoreFlow framework is introduced for automated LLM agent workflow generation and optimization.</p><p>&#8594; It employs code as a representation for workflows, enabling flexible and robust searches.</p><p>&#8594; Operators are used as predefined, reusable agent combinations, customizable by the generator.</p><p>&#8594; ScoreFlow uses an open-source LLM as the base model for workflow generation, minimizing costs.</p><p>&#8594; ScoreFlow optimizes workflow generators using preference data derived from evaluation scores.</p><p>&#8594; Score-DPO, a variant of Direct Preference Optimization, is proposed.</p><p>&#8594; Score-DPO incorporates quantitative evaluation scores directly into the preference optimization process.</p><p>&#8594; Score-DPO enhances sampling distribution by up-weighting preference pairs with larger score differences.</p><p>&#8594; Score-DPO incorporates evaluation scores into the Bradley-Terry ranking objective for improved learning.</p><p>-----</p><p>Key Insights &#128161;:</p><p>&#8594; ScoreFlow achieves high performance, scalability, and adaptability in LLM agent workflow optimization.</p><p>&#8594; Loss-gradient optimization in ScoreFlow offers more flexibility and scalability compared to discrete methods.</p><p>&#8594; Score-DPO effectively addresses inaccuracies in evaluation scores, improving optimization convergence.</p><p>&#8594; Adaptive workflow generation in ScoreFlow allows for task-specific operator selection and workflow complexity.</p><p>&#8594; ScoreFlow demonstrates robustness across different LLM architectures for both generators and executors.</p><p>&#8594; ScoreFlow enables smaller models to outperform larger models with better cost efficiency.</p><p>-----</p><p>Results &#128202;:</p><p>&#8594; ScoreFlow achieves an 8.2% average performance improvement over baselines across six benchmarks.</p><p>&#8594; ScoreFlow outperforms automated workflow optimization baselines like ADAS and Aflow.</p><p>&#8594; ScoreFlow, using Score-DPO, outperforms baselines like SFT, PPO and standard DPO in workflow optimization.</p><p>&#8594; ScoreFlow shows better scalability and performance advantage over Aflow on diverse combined datasets.</p>]]></content:encoded></item><item><title><![CDATA["Scaling up Test-Time Compute with Latent Reasoning: A Recurrent Depth Approach"]]></title><description><![CDATA[Below podcast on this paper is generated with Google's Illuminate.]]></description><link>https://www.rohan-paul.com/p/scaling-up-test-time-compute-with</link><guid isPermaLink="false">https://www.rohan-paul.com/p/scaling-up-test-time-compute-with</guid><dc:creator><![CDATA[Rohan Paul]]></dc:creator><pubDate>Sun, 16 Feb 2025 14:25:52 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!jNGx!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e28ff04-d31e-4463-b1ce-54634ffa7eec_1101x822.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!jNGx!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e28ff04-d31e-4463-b1ce-54634ffa7eec_1101x822.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!jNGx!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e28ff04-d31e-4463-b1ce-54634ffa7eec_1101x822.png 424w, https://substackcdn.com/image/fetch/$s_!jNGx!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e28ff04-d31e-4463-b1ce-54634ffa7eec_1101x822.png 848w, https://substackcdn.com/image/fetch/$s_!jNGx!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e28ff04-d31e-4463-b1ce-54634ffa7eec_1101x822.png 1272w, https://substackcdn.com/image/fetch/$s_!jNGx!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e28ff04-d31e-4463-b1ce-54634ffa7eec_1101x822.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!jNGx!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e28ff04-d31e-4463-b1ce-54634ffa7eec_1101x822.png" width="1101" height="822" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/7e28ff04-d31e-4463-b1ce-54634ffa7eec_1101x822.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:822,&quot;width&quot;:1101,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:280893,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!jNGx!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e28ff04-d31e-4463-b1ce-54634ffa7eec_1101x822.png 424w, https://substackcdn.com/image/fetch/$s_!jNGx!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e28ff04-d31e-4463-b1ce-54634ffa7eec_1101x822.png 848w, https://substackcdn.com/image/fetch/$s_!jNGx!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e28ff04-d31e-4463-b1ce-54634ffa7eec_1101x822.png 1272w, https://substackcdn.com/image/fetch/$s_!jNGx!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7e28ff04-d31e-4463-b1ce-54634ffa7eec_1101x822.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><a href="https://arxiv.org/abs/2502.05171">https://arxiv.org/abs/2502.05171</a></p><p>The paper addresses the challenge of enhancing reasoning in LLMs without drastically increasing model size or relying on specialized training data like chain-of-thought examples. Current methods often scale compute by generating more tokens, which can be inefficient.</p><p>This paper proposes a novel LLM architecture using a recurrent depth approach. It enables scaling test-time computation by iteratively refining latent representations, rather than verbalizing intermediate thoughts. This method works with standard training data and smaller context windows, potentially capturing reasoning forms beyond verbalization.</p><p>-----</p><p>&#128204; Recurrent architecture enables test-time compute scaling. This offers a parameter-efficient way to enhance reasoning without increasing model size. Deeper inference at test time improves performance on complex tasks.</p><p>&#128204; Latent space recurrence allows for implicit reasoning. This contrasts with explicit chain-of-thought. Model "thinks" in continuous space, potentially capturing non-verbal reasoning.</p><p>&#128204; Variable recurrence during training is key. Random iteration counts force model to generalize across compute budgets. Truncated backpropagation enables efficient training despite recurrent depth.</p><p>----------</p><p>Methods Explored in this Paper &#128295;:</p><p>&#8594; Introduces a depth-recurrent LLM architecture. It is built upon standard decoder-only transformer blocks.</p><p>&#8594; These blocks are organized into three parts: Prelude, Recurrent block, and Coda. The Prelude embeds input into a latent space.</p><p>&#8594; The core Recurrent block iteratively refines a latent state. The Coda un-embeds the latent state and predicts the next token.</p><p>&#8594; A key feature is the recurrent block which loops for a variable number of iterations during training and testing. This allows for scaling compute at test-time without increasing model parameters.</p><p>&#8594; During training, the number of recurrent iterations is randomly sampled from a log-normal Poisson distribution. This ensures the model learns to function with varying compute.</p><p>&#8594; To manage computational cost, truncated backpropagation is used. Backpropagation is limited to the last 8 recurrent iterations.</p><p>&#8594; The model is trained on a diverse dataset favoring code and mathematical reasoning data. This aims to promote emergent reasoning abilities.</p><p>-----</p><p>Key Insights &#128161;:</p><p>&#8594; Depth-recurrent models can effectively learn and improve performance by scaling test-time computation.</p><p>&#8594; Latent reasoning can capture complex reasoning beyond verbalization, like spatial or numerical reasoning.</p><p>&#8594; Recurrent models naturally support features like per-token adaptive compute, KV-cache sharing, and self-speculative decoding, simplifying LLM use cases.</p><p>&#8594; The model exhibits context-dependent convergence and path independence in its latent space, suggesting complex computational behaviors emerge with scale.</p><p>-----</p><p>Results &#128202;:</p><p>&#8594; The 3.5B parameter recurrent model achieves performance comparable to 7B parameter models on standard benchmarks.</p><p>&#8594; On mathematical reasoning tasks like GSM8K, the model surpasses most open-source models, showing significant gains from recurrent depth.</p><p>&#8594; Performance on harder tasks like ARC challenge and GSM8K improves significantly with increased test-time compute (more recurrent iterations).</p><p>&#8594; At 180B training tokens, the recurrent model outperforms a non-recurrent baseline, especially on challenging reasoning tasks.</p>]]></content:encoded></item><item><title><![CDATA["Scaling Laws in Patchification: An Image Is Worth 50,176 Tokens And More"]]></title><description><![CDATA[Below podcast on this paper is generated with Google's Illuminate.]]></description><link>https://www.rohan-paul.com/p/scaling-laws-in-patchification-an</link><guid isPermaLink="false">https://www.rohan-paul.com/p/scaling-laws-in-patchification-an</guid><dc:creator><![CDATA[Rohan Paul]]></dc:creator><pubDate>Sun, 16 Feb 2025 14:21:36 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!n9LT!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb16118ac-033b-42a3-84b4-4a8971b89067_924x866.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!n9LT!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb16118ac-033b-42a3-84b4-4a8971b89067_924x866.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!n9LT!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb16118ac-033b-42a3-84b4-4a8971b89067_924x866.png 424w, https://substackcdn.com/image/fetch/$s_!n9LT!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb16118ac-033b-42a3-84b4-4a8971b89067_924x866.png 848w, https://substackcdn.com/image/fetch/$s_!n9LT!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb16118ac-033b-42a3-84b4-4a8971b89067_924x866.png 1272w, https://substackcdn.com/image/fetch/$s_!n9LT!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb16118ac-033b-42a3-84b4-4a8971b89067_924x866.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!n9LT!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb16118ac-033b-42a3-84b4-4a8971b89067_924x866.png" width="924" height="866" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b16118ac-033b-42a3-84b4-4a8971b89067_924x866.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:866,&quot;width&quot;:924,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:391070,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!n9LT!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb16118ac-033b-42a3-84b4-4a8971b89067_924x866.png 424w, https://substackcdn.com/image/fetch/$s_!n9LT!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb16118ac-033b-42a3-84b4-4a8971b89067_924x866.png 848w, https://substackcdn.com/image/fetch/$s_!n9LT!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb16118ac-033b-42a3-84b4-4a8971b89067_924x866.png 1272w, https://substackcdn.com/image/fetch/$s_!n9LT!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb16118ac-033b-42a3-84b4-4a8971b89067_924x866.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><a href="https://arxiv.org/abs/2502.03738">https://arxiv.org/abs/2502.03738</a></p><p>The common approach in Vision Transformer models compresses images into patches, reducing computational cost but potentially losing visual information. This paper addresses the problem of information loss due to image patchification in vision models.</p><p>This paper proposes to reduce the patch size, even to single pixels, to minimize information loss and improve model performance.</p><p>-----</p><p>&#128204; Smaller patch sizes in Vision Transformers access finer image details. This enhances visual information fidelity directly at the input level. Models benefit from richer representations without complex architectural changes.</p><p>&#128204; Computational cost shifts from model parameters to sequence length. Mamba architecture's linear complexity becomes crucial. Pixel-level tokenization is now feasible for high resolution images, unlocking new scaling dimensions.</p><p>&#128204; Decoder heads become less critical for tasks like semantic segmentation. High fidelity encoders with pixel tokens suffice. This simplifies architecture and suggests a move towards encoder-only visual models.</p><p>----------</p><p>Methods Explored in this Paper &#128295;:</p><p>&#8594; The paper investigates the impact of patch size by conducting experiments with Vision Transformer and Mamba based architectures.</p><p>&#8594; They systematically reduced the patch size from 16x16 down to 1x1 pixel.</p><p>&#8594; This reduction in patch size increases the input sequence length significantly, up to 50,176 tokens for a 224x224 image.</p><p>&#8594; Experiments were performed on ImageNet-1k classification, ADE20k semantic segmentation, and COCO object detection/instance segmentation tasks.</p><p>&#8594; Both Vision Transformer and Adventurer architectures were used to ensure findings are generalizable.</p><p>-----</p><p>Key Insights &#128161;:</p><p>&#8594; A scaling law in patchification is observed.</p><p>&#8594; Smaller patch sizes consistently improve model performance across various vision tasks, input resolutions, and architectures.</p><p>&#8594; Patchification is identified as a compromise for computational efficiency, not a necessity for effective vision models.</p><p>&#8594; Reducing patch size unlocks crucial visual information previously lost through compression.</p><p>&#8594; Task specific decoder heads become less critical for dense prediction tasks when using smaller patch sizes.</p><p>-----</p><p>Results &#128202;:</p><p>&#8594; ImageNet-1k classification accuracy improved from 82.6% to 84.6% on Adventurer-Base model by reducing patch size to 1x1 on 224x224 images.</p><p>&#8594; On ADE20k semantic segmentation, mIoU improved consistently as patch size decreased, reaching 46.8% mIoU with Adventurer-Base and 2x2 patch size without a decoder head.</p><p>&#8594; COCO object detection APb improved from 44.7% to 48.7% on Adventurer-Tiny and from 44.1% to 50.3% on Adventurer-Base by reducing patch size.</p>]]></content:encoded></item><item><title><![CDATA["ReasonFlux: Hierarchical LLM Reasoning via Scaling Thought Templates"]]></title><description><![CDATA[Below podcast on this paper is generated with Google's Illuminate.]]></description><link>https://www.rohan-paul.com/p/reasonflux-hierarchical-llm-reasoning</link><guid isPermaLink="false">https://www.rohan-paul.com/p/reasonflux-hierarchical-llm-reasoning</guid><dc:creator><![CDATA[Rohan Paul]]></dc:creator><pubDate>Sun, 16 Feb 2025 14:15:07 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!4mCN!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F71383d5b-ead1-475b-a588-11ebc6d9abf1_1297x821.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!4mCN!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F71383d5b-ead1-475b-a588-11ebc6d9abf1_1297x821.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!4mCN!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F71383d5b-ead1-475b-a588-11ebc6d9abf1_1297x821.png 424w, https://substackcdn.com/image/fetch/$s_!4mCN!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F71383d5b-ead1-475b-a588-11ebc6d9abf1_1297x821.png 848w, https://substackcdn.com/image/fetch/$s_!4mCN!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F71383d5b-ead1-475b-a588-11ebc6d9abf1_1297x821.png 1272w, https://substackcdn.com/image/fetch/$s_!4mCN!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F71383d5b-ead1-475b-a588-11ebc6d9abf1_1297x821.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!4mCN!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F71383d5b-ead1-475b-a588-11ebc6d9abf1_1297x821.png" width="1297" height="821" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/71383d5b-ead1-475b-a588-11ebc6d9abf1_1297x821.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:821,&quot;width&quot;:1297,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:296212,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!4mCN!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F71383d5b-ead1-475b-a588-11ebc6d9abf1_1297x821.png 424w, https://substackcdn.com/image/fetch/$s_!4mCN!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F71383d5b-ead1-475b-a588-11ebc6d9abf1_1297x821.png 848w, https://substackcdn.com/image/fetch/$s_!4mCN!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F71383d5b-ead1-475b-a588-11ebc6d9abf1_1297x821.png 1272w, https://substackcdn.com/image/fetch/$s_!4mCN!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F71383d5b-ead1-475b-a588-11ebc6d9abf1_1297x821.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><a href="https://arxiv.org/abs/2502.06772">https://arxiv.org/abs/2502.06772</a></p><p>The challenge lies in enhancing LLMs' (LLMs) complex reasoning, particularly in tasks like math problem-solving, which demands extensive search and fine-grained thinking. Current methods struggle with efficiency and generalization.</p><p>This paper introduces ReasonFlux. It addresses these limitations through hierarchical LLM reasoning. ReasonFlux uses scaled thought templates to optimize the reasoning process.</p><p>-----</p><p>&#128204; ReasonFlux tackles complex reasoning via structured thought templates. This is a form of knowledge distillation. It encodes expert problem-solving strategies into reusable modules, significantly boosting mathematical reasoning accuracy for LLMs.</p><p>&#128204; Hierarchical reinforcement learning on template trajectories is a key innovation. It moves beyond step-by-step Chain of Thought. By optimizing template sequences, ReasonFlux learns effective high-level reasoning plans, improving generalization.</p><p>&#128204; ReasonFlux's adaptive template scaling at inference enables efficient problem exploration. This dynamic approach contrasts with static inference methods. It achieves a better exploration-exploitation trade-off, leading to superior performance on challenging benchmarks.</p><p>----------</p><p>Methods Explored in this Paper &#128295;:</p><p>&#8594; ReasonFlux employs a structured thought template library. This library contains around 500 high-level templates. These templates are designed for efficient retrieval and adaptation to various reasoning problems.</p><p>&#8594; Hierarchical reinforcement learning is used. It optimizes a base LLM to create optimal template trajectories. This is done on sequences of thought templates, not raw Chain of Thought data. This helps in planning solutions for complex problems step-by-step using templates.</p><p>&#8594; A new inference scaling system is introduced. This system adaptively scales thought templates during inference. It enables hierarchical LLM reasoning by dynamically selecting and applying templates. This adaptive scaling aims to balance exploration and exploitation in the reasoning process.</p><p>-----</p><p>Key Insights &#128161;:</p><p>&#8594; ReasonFlux simplifies complex reasoning by using hierarchical template-based approach. It moves from searching in the vast original problem space to a more manageable template space.</p><p>&#8594; The structured template library is crucial for efficient retrieval of relevant knowledge. Templates are designed to be reusable and generalizable across similar problems.</p><p>&#8594; Hierarchical reinforcement learning allows the model to learn effective strategies for combining and sequencing thought templates. This leads to better planning of reasoning paths.</p><p>&#8594; Adaptive template scaling at inference allows for dynamic problem-solving. ReasonFlux adjusts its approach based on the problem's complexity, improving efficiency and accuracy.</p><p>-----</p><p>Results &#128202;:</p><p>&#8594; ReasonFlux-32B achieves 91.2% accuracy on the MATH benchmark. This surpasses OpenAI o1-preview by 6.7%.</p><p>&#8594; On the AIME 2024 benchmark, ReasonFlux-32B achieves 56.7% accuracy. This outperforms o1-preview by 27% and DeepSeek-V3 by 45%.</p><p>&#8594; ReasonFlux-32B achieves 63.3% accuracy on OlympiadBench, exceeding DeepSeek-V3 by 14%.</p>]]></content:encoded></item><item><title><![CDATA["QuEST: Stable Training of LLMs with 1-Bit Weights and Activations"]]></title><description><![CDATA[Below podcast on this paper is generated with Google's Illuminate.]]></description><link>https://www.rohan-paul.com/p/quest-stable-training-of-llms-with</link><guid isPermaLink="false">https://www.rohan-paul.com/p/quest-stable-training-of-llms-with</guid><dc:creator><![CDATA[Rohan Paul]]></dc:creator><pubDate>Sun, 16 Feb 2025 14:10:48 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!aeAf!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f3013d3-ca30-4ebf-a185-5a351167aab1_845x823.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!aeAf!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f3013d3-ca30-4ebf-a185-5a351167aab1_845x823.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!aeAf!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f3013d3-ca30-4ebf-a185-5a351167aab1_845x823.png 424w, https://substackcdn.com/image/fetch/$s_!aeAf!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f3013d3-ca30-4ebf-a185-5a351167aab1_845x823.png 848w, https://substackcdn.com/image/fetch/$s_!aeAf!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f3013d3-ca30-4ebf-a185-5a351167aab1_845x823.png 1272w, https://substackcdn.com/image/fetch/$s_!aeAf!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f3013d3-ca30-4ebf-a185-5a351167aab1_845x823.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!aeAf!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f3013d3-ca30-4ebf-a185-5a351167aab1_845x823.png" width="845" height="823" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/0f3013d3-ca30-4ebf-a185-5a351167aab1_845x823.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:823,&quot;width&quot;:845,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:354955,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!aeAf!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f3013d3-ca30-4ebf-a185-5a351167aab1_845x823.png 424w, https://substackcdn.com/image/fetch/$s_!aeAf!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f3013d3-ca30-4ebf-a185-5a351167aab1_845x823.png 848w, https://substackcdn.com/image/fetch/$s_!aeAf!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f3013d3-ca30-4ebf-a185-5a351167aab1_845x823.png 1272w, https://substackcdn.com/image/fetch/$s_!aeAf!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0f3013d3-ca30-4ebf-a185-5a351167aab1_845x823.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><a href="https://arxiv.org/abs/2502.05003">https://arxiv.org/abs/2502.05003</a></p><p>The challenge of high computational costs in LLMs restricts their efficiency. Quantization Aware Training struggles to maintain accuracy at very low bit-widths, with 8-bit quantization often seen as the limit for practical training.</p><p>This paper introduces QuEST, a novel Quantization Aware Training method. QuEST achieves stable training of LLMs using extremely low 1-bit weights and activations. QuEST also pushes the Pareto frontier to 4-bit quantization, surpassing the accuracy of standard 16-bit formats at comparable model sizes.</p><p>-----</p><p>&#128204; QuEST's Hadamard Transform preprocessing is a key innovation. It reshapes weight distributions for effective quantization. This technique unlocks surprisingly stable and accurate training even at 1-bit, pushing limits of low-precision Quantization Aware Training.</p><p>&#128204; The Trust Gradient Estimator in QuEST directly tackles the core challenge of noisy gradients in low-bit training. By selectively trusting gradients based on quantization error, it stabilizes optimization. This is crucial for 1-bit training viability.</p><p>&#128204; QuEST practically demonstrates that 4-bit LLMs can surpass Brain Float 16 (BF16) performance. This redefines the Pareto frontier for efficient LLM training. It immediately enables smaller, faster, and cheaper high-accuracy models.</p><p>----------</p><p>Methods Explored in this Paper &#128295;:</p><p>&#8594; QuEST enhances Quantization Aware Training through two core innovations.</p><p>&#8594; It introduces a novel approach to distribution fitting in the forward pass. This involves applying a Hadamard Transform for normalization before quantizing weights and activations. This Hadamard Transform helps to shape the distribution of weights to be closer to Gaussian, which is more suitable for quantization.</p><p>&#8594; QuEST employs Mean Squared Error optimal fitting to determine the best quantization grid for the transformed distributions. This ensures accurate and fast quantization.</p><p>&#8594; For the backward pass, QuEST presents a Trust Gradient Estimator. This estimator minimizes the discrepancy between gradients calculated with quantized parameters and full-precision gradients.</p><p>&#8594; The Trust Gradient Estimator selectively reduces the influence of gradients from weight components that have high quantization errors in the forward pass. This strategy stabilizes training by mitigating the impact of outlier quantization errors on gradient updates.</p><p>-----</p><p>Key Insights &#128161;:</p><p>&#8594; QuEST enables stable training of LLMs with just 1-bit weights and activations. This was not previously considered feasible with standard Quantization Aware Training methods.</p><p>&#8594; Models trained with QuEST using 4-bit weights and activations achieve superior accuracy compared to models using Brain Float 16 (BF16) format, especially when considering model size and inference cost. QuEST shifts the Pareto-optimal frontier for LLM training to lower bit-widths.</p><p>&#8594; Hadamard Transform preprocessing significantly improves the effectiveness of the trust estimation method, contributing to stable training and better performance.</p><p>&#8594; Analysis of efficiency factors reveals that 4-bit quantization offers the optimal balance between model size, computational cost, and accuracy in overtraining scenarios.</p><p>-----</p><p>Results &#128202;:</p><p>&#8594; QuEST trained models with 4-bit weights and activations outperform BF16 models of nearly 4 times larger size in terms of accuracy, as shown in Figure 1.</p><p>&#8594; QuEST demonstrates lower perplexity than LSQ (Learned Step Size Quantization), especially at lower bit-widths, as shown in Figure 3. For instance, at W4A4 configuration, QuEST achieves 29.08 perplexity, while LSQ reaches 30.27.</p><p>&#8594; GPU kernel implementation of QuEST achieves per-layer inference speedups from 1.2&#215; to 2.4&#215; for an 800M parameter model and 2.3&#215; to 3.9&#215; for a 7B parameter model compared to BF16, as depicted in Figure 6.</p><p>&#8594; End-to-end inference speedup of 1.3&#215; to 1.5&#215; is observed for QuEST INT4 over BF16 in prefill stage using an 800M parameter model on RTX 4090, as shown in Figure 7.</p><p>&#8594; Zero-shot evaluation on HellaSWAG benchmark shows comparable accuracy between QuEST 4-bit model (39.22%) and BF16 model (39.52%) of 800M parameters, as indicated in Table 2, suggesting near lossless quantization.</p>]]></content:encoded></item><item><title><![CDATA["QLIP: Text-Aligned Visual Tokenization Unifies Auto-Regressive Multimodal Understanding and Generation"]]></title><description><![CDATA[Below podcast on this paper is generated with Google's Illuminate.]]></description><link>https://www.rohan-paul.com/p/qlip-text-aligned-visual-tokenization</link><guid isPermaLink="false">https://www.rohan-paul.com/p/qlip-text-aligned-visual-tokenization</guid><dc:creator><![CDATA[Rohan Paul]]></dc:creator><pubDate>Sun, 16 Feb 2025 10:29:19 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!5u1u!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F314e305c-8ec8-4878-b9f0-41de8d883eee_1076x857.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!5u1u!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F314e305c-8ec8-4878-b9f0-41de8d883eee_1076x857.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!5u1u!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F314e305c-8ec8-4878-b9f0-41de8d883eee_1076x857.png 424w, https://substackcdn.com/image/fetch/$s_!5u1u!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F314e305c-8ec8-4878-b9f0-41de8d883eee_1076x857.png 848w, https://substackcdn.com/image/fetch/$s_!5u1u!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F314e305c-8ec8-4878-b9f0-41de8d883eee_1076x857.png 1272w, https://substackcdn.com/image/fetch/$s_!5u1u!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F314e305c-8ec8-4878-b9f0-41de8d883eee_1076x857.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!5u1u!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F314e305c-8ec8-4878-b9f0-41de8d883eee_1076x857.png" width="1076" height="857" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/314e305c-8ec8-4878-b9f0-41de8d883eee_1076x857.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:857,&quot;width&quot;:1076,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:290315,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!5u1u!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F314e305c-8ec8-4878-b9f0-41de8d883eee_1076x857.png 424w, https://substackcdn.com/image/fetch/$s_!5u1u!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F314e305c-8ec8-4878-b9f0-41de8d883eee_1076x857.png 848w, https://substackcdn.com/image/fetch/$s_!5u1u!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F314e305c-8ec8-4878-b9f0-41de8d883eee_1076x857.png 1272w, https://substackcdn.com/image/fetch/$s_!5u1u!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F314e305c-8ec8-4878-b9f0-41de8d883eee_1076x857.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><a href="https://arxiv.org/abs/2502.05178">https://arxiv.org/abs/2502.05178</a></p><p>The challenge is to create a unified model for both visual understanding and generation, but current visual tokenizers are optimized for only one of these tasks. This paper introduces Quantized Language-Image Pretraining (QLIP) to bridge this gap.</p><p>This paper proposes QLIP, a visual tokenizer trained with both reconstruction and language-image alignment objectives. QLIP uses a two-stage training process and dynamic loss weighting to achieve state-of-the-art performance in both areas.</p><p>-----</p><p>&#128204; QLIP pioneers text-aligned visual tokenization. It uniquely integrates visual reconstruction with semantic understanding within the tokenizer itself. This enables superior zero-shot image classification at 74.3% accuracy and improved generation FID at 15.29.</p><p>&#128204; The two-stage training of QLIP is crucial. Stage one balances contrastive alignment and reconstruction using dynamic loss weights. Stage two optimizes reconstruction with perceptual and adversarial losses. This overcomes memory limits and gradient conflicts.</p><p>&#128204; QLIP's unified multimodal model, UM3, demonstrates versatility. It handles text-only, image-to-text, and text-to-image tasks within a single architecture. UM3 achieves comparable performance to specialized models, showcasing QLIP's effectiveness.</p><p>----------</p><p>Methods Explored in this Paper &#128295;:</p><p>&#8594; QLIP is a visual tokenizer based on Binary Spherical Quantization. It is trained as an autoencoder to reconstruct images into discrete visual tokens.</p><p>&#8594; Simultaneously, QLIP is trained with a contrastive language-image objective. This aligns visual tokens with text embeddings, enhancing semantic representation.</p><p>&#8594; A two-stage training pipeline is employed. Stage 1 trains QLIP with both alignment and reconstruction losses using a memory-efficient Transformer. Stage 2 fine-tunes the quantizer and decoder with reconstruction losses, after freezing the visual encoder and dropping the text encoder. This improves reconstruction quality by using perceptual and adversarial losses without memory limitations.</p><p>&#8594; To balance the competing alignment and reconstruction objectives, an automated weighting scheme is introduced. Loss terms are weighted by the inverse of their post-hoc loss values. This addresses the differing gradient magnitudes and convergence rates of the two objectives.</p><p>-----</p><p>Key Insights &#128161;:</p><p>&#8594; Balancing reconstruction and alignment objectives during visual tokenizer training is crucial for achieving strong performance in both understanding and generation.</p><p>&#8594; A two-stage training process effectively addresses the challenges of large batch contrastive learning and memory-intensive reconstruction losses.</p><p>&#8594; Initializing the visual encoder with Masked Image Modeling or CLIP pre-training significantly accelerates convergence and improves performance.</p><p>-----</p><p>Results &#128202;:</p><p>&#8594; QLIP-B achieves 74.3% zero-shot classification accuracy on ImageNet, comparable to CLIP models while also providing visual tokenization.</p><p>&#8594; QLIP-B achieves a reconstruction rFID of 3.21, demonstrating state-of-the-art reconstruction quality for a semantically aligned tokenizer.</p><p>&#8594; In text-conditioned image generation on MS-COCO 30k, LlamaGen with QLIP-B achieves a generation FID of 15.29, outperforming LlamaGen with VQGAN (15.68).</p>]]></content:encoded></item><item><title><![CDATA["PlotGen: Multi-Agent LLM-based Scientific Data Visualization via Multimodal Feedback"]]></title><description><![CDATA[Below podcast on this paper is generated with Google's Illuminate.]]></description><link>https://www.rohan-paul.com/p/plotgen-multi-agent-llm-based-scientific</link><guid isPermaLink="false">https://www.rohan-paul.com/p/plotgen-multi-agent-llm-based-scientific</guid><dc:creator><![CDATA[Rohan Paul]]></dc:creator><pubDate>Sun, 16 Feb 2025 10:25:29 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!ZHad!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F536496ed-2e4b-4870-9367-bb3aeb134a7b_1090x769.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!ZHad!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F536496ed-2e4b-4870-9367-bb3aeb134a7b_1090x769.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!ZHad!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F536496ed-2e4b-4870-9367-bb3aeb134a7b_1090x769.png 424w, https://substackcdn.com/image/fetch/$s_!ZHad!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F536496ed-2e4b-4870-9367-bb3aeb134a7b_1090x769.png 848w, https://substackcdn.com/image/fetch/$s_!ZHad!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F536496ed-2e4b-4870-9367-bb3aeb134a7b_1090x769.png 1272w, https://substackcdn.com/image/fetch/$s_!ZHad!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F536496ed-2e4b-4870-9367-bb3aeb134a7b_1090x769.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!ZHad!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F536496ed-2e4b-4870-9367-bb3aeb134a7b_1090x769.png" width="1090" height="769" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/536496ed-2e4b-4870-9367-bb3aeb134a7b_1090x769.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:769,&quot;width&quot;:1090,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:360731,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!ZHad!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F536496ed-2e4b-4870-9367-bb3aeb134a7b_1090x769.png 424w, https://substackcdn.com/image/fetch/$s_!ZHad!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F536496ed-2e4b-4870-9367-bb3aeb134a7b_1090x769.png 848w, https://substackcdn.com/image/fetch/$s_!ZHad!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F536496ed-2e4b-4870-9367-bb3aeb134a7b_1090x769.png 1272w, https://substackcdn.com/image/fetch/$s_!ZHad!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F536496ed-2e4b-4870-9367-bb3aeb134a7b_1090x769.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><a href="https://arxiv.org/abs/2502.00988">https://arxiv.org/abs/2502.00988</a></p><p>The creation of scientific data visualizations is challenging for novice users due to the complexity of tools and visualization techniques. Existing LLMs struggle with accuracy in generating visualization code and require iterative debugging.</p><p>This paper introduces PlotGen, a multi-agent framework that utilizes multimodal feedback to iteratively refine and enhance the accuracy of scientific visualizations generated by LLMs.</p><p>-----</p><p>&#128204; PlotGen effectively decomposes the complex visualization task into modular agents. Each agent specializes in query planning, code generation, or multimodal feedback. This division of labor enables targeted error correction and enhances overall accuracy.</p><p>&#128204; Multimodal feedback is a key innovation. PlotGen uses visual, lexical, and numerical feedback agents. These agents mimic human sensory input to iteratively refine plots generated by LLMs, addressing inherent code generation inaccuracies.</p><p>&#128204; PlotGen improves accessibility for novice users. By automating debugging through self-reflection and multimodal feedback, it lowers the technical barrier to creating accurate scientific visualizations, enhancing user productivity.</p><p>----------</p><p>Methods Explored in this Paper &#128295;:</p><p>&#8594; PlotGen employs a multi-agent framework to automate scientific data visualization.</p><p>&#8594; It starts with a Query Planning Agent. This agent breaks down complex user requests into a sequence of executable steps using chain-of-thought prompting.</p><p>&#8594; A Code Generation Agent then converts these steps into executable Python code for plotting. This agent also includes a self-debugging mechanism to handle code execution errors.</p><p>&#8594; PlotGen incorporates three feedback agents to refine the visualizations. These are Numeric, Lexical, and Visual Feedback Agents.</p><p>&#8594; The Numeric Feedback Agent ensures data accuracy and correct plot type by comparing de-rendered data from the plot with the original data.</p><p>&#8594; The Lexical Feedback Agent verifies the accuracy of textual elements like titles and labels by comparing them to user requirements and data.</p><p>&#8594; The Visual Feedback Agent assesses visual aspects such as color schemes and layout to ensure alignment with user specifications.</p><p>&#8594; These feedback agents provide iterative feedback to the Code Generation Agent for self-reflection and refinement of the generated plots.</p><p>-----</p><p>Key Insights &#128161;:</p><p>&#8594; Multimodal feedback is crucial for enhancing the accuracy of LLM generated scientific visualizations.</p><p>&#8594; PlotGen leverages visual, lexical, and numerical feedback to rectify errors in data, text, and visual aesthetics, through self-reflection.</p><p>&#8594; This multi-agent approach significantly improves the quality and trustworthiness of LLM generated plots for scientific data visualization.</p><p>-----</p><p>Results &#128202;:</p><p>&#8594; PlotGen outperforms baseline methods such as Direct Decoding, Zero-Shot Chain-of-thought, and MatPlotAgent on the MatPlotBench dataset.</p><p>&#8594; PlotGen achieves a 4-6% performance improvement over these strong baselines across various LLM configurations.</p><p>&#8594; User evaluations indicate increased user trust in PlotGen generated visualizations and a reduction in debugging time for novice users.</p><p></p>]]></content:encoded></item><item><title><![CDATA["PILAF: Optimal Human Preference Sampling for Reward Modeling"]]></title><description><![CDATA[Below podcast on this paper is generated with Google's Illuminate.]]></description><link>https://www.rohan-paul.com/p/pilaf-optimal-human-preference-sampling</link><guid isPermaLink="false">https://www.rohan-paul.com/p/pilaf-optimal-human-preference-sampling</guid><dc:creator><![CDATA[Rohan Paul]]></dc:creator><pubDate>Sun, 16 Feb 2025 10:20:04 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!B3kc!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F55b4068b-ef8b-43d9-adf4-71a7838fd697_1102x792.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!B3kc!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F55b4068b-ef8b-43d9-adf4-71a7838fd697_1102x792.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!B3kc!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F55b4068b-ef8b-43d9-adf4-71a7838fd697_1102x792.png 424w, https://substackcdn.com/image/fetch/$s_!B3kc!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F55b4068b-ef8b-43d9-adf4-71a7838fd697_1102x792.png 848w, https://substackcdn.com/image/fetch/$s_!B3kc!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F55b4068b-ef8b-43d9-adf4-71a7838fd697_1102x792.png 1272w, https://substackcdn.com/image/fetch/$s_!B3kc!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F55b4068b-ef8b-43d9-adf4-71a7838fd697_1102x792.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!B3kc!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F55b4068b-ef8b-43d9-adf4-71a7838fd697_1102x792.png" width="1102" height="792" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/55b4068b-ef8b-43d9-adf4-71a7838fd697_1102x792.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:792,&quot;width&quot;:1102,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:194962,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!B3kc!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F55b4068b-ef8b-43d9-adf4-71a7838fd697_1102x792.png 424w, https://substackcdn.com/image/fetch/$s_!B3kc!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F55b4068b-ef8b-43d9-adf4-71a7838fd697_1102x792.png 848w, https://substackcdn.com/image/fetch/$s_!B3kc!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F55b4068b-ef8b-43d9-adf4-71a7838fd697_1102x792.png 1272w, https://substackcdn.com/image/fetch/$s_!B3kc!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F55b4068b-ef8b-43d9-adf4-71a7838fd697_1102x792.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><a href="https://arxiv.org/abs/2502.04270">https://arxiv.org/abs/2502.04270</a></p><p>The challenge in Reinforcement Learning from Human Feedback (RLHF) is that current methods for collecting preference data for reward modeling are inefficient and do not directly optimize for true human values. This paper addresses the misalignment between reward model training and maximizing the actual human preference objective in RLHF.</p><p>This paper proposes Policy-Interpolated Learning for Aligned Feedback (PILAF). PILAF is a novel sampling strategy for preference labeling. It aligns preference learning with the goal of maximizing the underlying true human reward.</p><p>-----</p><p>&#128204; PILAF directly tackles the objective misalignment in Reinforcement Learning from Human Feedback. It aligns reward model training with the true preference objective by shaping the preference sampling, enhancing learning efficiency.</p><p>&#128204; Statistically, PILAF strategically samples data in directions of maximal objective sensitivity. This reduces variance in preference data, leading to more robust and data-efficient reward model training.</p><p>&#128204; PILAF's practical strength lies in its simple implementation and lack of hyperparameters. This allows for immediate improvement in existing Direct Preference Optimization pipelines with minimal overhead.</p><p>----------</p><p>Methods Explored in this Paper &#128295;:</p><p>&#8594; Introduces Theoretically Grounded Policy-Interpolated Learning for Aligned Feedback (T-PILAF). T-PILAF generates response pairs by interpolating between the current policy and a reference policy. This balances exploration and exploitation during data collection.</p><p>&#8594; T-PILAF uses two modified policies, &#960;+ and &#960;&#8722;, derived from the current policy &#960;. &#960;+ encourages exploration in areas favored by &#960;, and &#960;&#8722; explores less favored areas.</p><p>&#8594; PILAF is presented as a practical simplification of T-PILAF. PILAF avoids complex normalization factors and approximates policy interpolation token-wise.</p><p>&#8594; PILAF samples responses from either the current policy &#960;, or by interpolating logits of the current policy &#960; and reference policy &#960;_ref. This interpolation uses the KL regularization coefficient &#946;.</p><p>-----</p><p>Key Insights &#128161;:</p><p>&#8594; T-PILAF aligns the gradient of the reward model loss with the policy gradient of the true human preference objective. This alignment ensures that minimizing the reward model loss directly contributes to maximizing human values.</p><p>&#8594; T-PILAF improves statistical efficiency. It aligns optimization with directions of greatest sensitivity in the human preference objective. This leads to more informative preference data and reduces training variance.</p><p>&#8594; PILAF, derived from T-PILAF, inherits these theoretical benefits. It improves sample efficiency and performance in practical RLHF settings.</p><p>-----</p><p>Results &#128202;:</p><p>&#8594; In iterative Direct Preference Optimization (DPO), PILAF achieves baseline reward levels with 40% less training time.</p><p>&#8594; In online DPO, PILAF demonstrates a better Reward-KL trade-off. It achieves higher reward with lower KL divergence compared to Vanilla and Hybrid Sampling.</p><p>&#8594; In robustness tests with an overfitted initial model, PILAF escapes suboptimal regions and achieves higher reward and lower KL divergence than Vanilla Sampling.</p><p>&#8594; PILAF reduces annotation and computation costs by over 40% in iterative DPO while maintaining comparable performance.</p>]]></content:encoded></item><item><title><![CDATA["Partially Rewriting a Transformer in Natural Language"]]></title><description><![CDATA[Below podcast on this paper is generated with Google's Illuminate.]]></description><link>https://www.rohan-paul.com/p/partially-rewriting-a-transformer</link><guid isPermaLink="false">https://www.rohan-paul.com/p/partially-rewriting-a-transformer</guid><dc:creator><![CDATA[Rohan Paul]]></dc:creator><pubDate>Sun, 16 Feb 2025 10:09:03 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!wlNn!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70056808-391b-486c-9872-172903bda68e_1082x875.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!wlNn!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70056808-391b-486c-9872-172903bda68e_1082x875.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!wlNn!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70056808-391b-486c-9872-172903bda68e_1082x875.png 424w, https://substackcdn.com/image/fetch/$s_!wlNn!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70056808-391b-486c-9872-172903bda68e_1082x875.png 848w, https://substackcdn.com/image/fetch/$s_!wlNn!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70056808-391b-486c-9872-172903bda68e_1082x875.png 1272w, https://substackcdn.com/image/fetch/$s_!wlNn!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70056808-391b-486c-9872-172903bda68e_1082x875.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!wlNn!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70056808-391b-486c-9872-172903bda68e_1082x875.png" width="1082" height="875" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/70056808-391b-486c-9872-172903bda68e_1082x875.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:875,&quot;width&quot;:1082,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:358810,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!wlNn!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70056808-391b-486c-9872-172903bda68e_1082x875.png 424w, https://substackcdn.com/image/fetch/$s_!wlNn!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70056808-391b-486c-9872-172903bda68e_1082x875.png 848w, https://substackcdn.com/image/fetch/$s_!wlNn!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70056808-391b-486c-9872-172903bda68e_1082x875.png 1272w, https://substackcdn.com/image/fetch/$s_!wlNn!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70056808-391b-486c-9872-172903bda68e_1082x875.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><a href="https://arxiv.org/abs/2501.18838">https://arxiv.org/abs/2501.18838</a></p><p>The challenge is understanding the internal workings of LLMs. Current interpretability methods are limited. This paper explores rewriting parts of an LLM using natural language for better understanding.</p><p>This paper proposes a method to partially rewrite a Transformer layer with natural language. It uses sparse representations and LLMs to explain and simulate neuron activations, then integrates these back into the original model.</p><p>-----</p><p>&#128204; Sparse transcoders offer a path to interpretability by approximating neural networks with sparse, explainable features. Natural language descriptions, though imperfect, bridge the gap to human understanding.</p><p>&#128204; Quantile normalization is essential for calibrating LLM-predicted activations. Raw LLM outputs for neuron activity are poorly distributed, hindering direct integration without statistical correction.</p><p>&#128204; Current natural language explanations are not yet sufficiently precise to fully replace neural network components. The performance remains close to zero ablation, indicating a need for richer, more specific explanations.</p><p>----------</p><p>Methods Explored in this Paper &#128295;:</p><p>&#8594; The paper trains a sparse transcoder. The transcoder approximates a feedforward network within the LLM. It uses a TopK activation function to enforce sparsity.</p><p>&#8594; A skip connection is added to the transcoder architecture. This improves approximation without affecting interpretability. A sparse autoencoder (SAE) is also trained on the residual stream.</p><p>&#8594; Automated interpretability pipeline generates natural language explanations for transcoder and SAE latents. This pipeline is used to score the quality of these explanations.</p><p>&#8594; An LLM predicts latent activation. It uses the generated explanation and surrounding text context. This prediction is for a single token.</p><p>&#8594; Quantile normalization calibrates the LLM's activation predictions. This aligns the predicted activation distribution with the true activation distribution. This step is crucial for performance.</p><p>-----</p><p>Key Insights &#128161;:</p><p>&#8594; Current automatically generated natural language explanations are not sufficiently specific. This lack of specificity hinders performance when rewriting model components.</p><p>&#8594; Simply predicting whether a feature is active is insufficient. Explanations must also accurately identify contexts where a feature is <em>*not*</em> active. Specificity is as important as sensitivity.</p><p>&#8594; Quantile normalization significantly improves performance. It addresses the issue of low specificity in explanations by correcting falsely activated neurons.</p><p>-----</p><p>Results &#128202;:</p><p>&#8594; Rewriting the entire transcoder with natural language explanations results in a cross-entropy loss similar to a Pythia model trained on only 10-15% of the data. This is comparable to replacing the transcoder output with zeros.</p><p>&#8594; Randomly selecting latents for rewriting leads to worse performance than zeroing out the MLP.</p><p>&#8594; Using quantile normalization significantly improves performance compared to not using it. Normalization makes rewritten models perform better than zeroing out components.</p><p>&#8594; Detection scores correlate with explanation quality. Higher detection scores indicate more specific and sensitive explanations.</p>]]></content:encoded></item><item><title><![CDATA["On-device Sora: Enabling Diffusion-Based Text-to-Video Generation for Mobile Devices"]]></title><description><![CDATA[Below podcast on this paper is generated with Google's Illuminate.]]></description><link>https://www.rohan-paul.com/p/on-device-sora-enabling-diffusion</link><guid isPermaLink="false">https://www.rohan-paul.com/p/on-device-sora-enabling-diffusion</guid><dc:creator><![CDATA[Rohan Paul]]></dc:creator><pubDate>Sat, 15 Feb 2025 14:21:55 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!cIpN!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3914a985-e6fc-452d-8332-24e2c7f38ea5_888x873.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!cIpN!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3914a985-e6fc-452d-8332-24e2c7f38ea5_888x873.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!cIpN!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3914a985-e6fc-452d-8332-24e2c7f38ea5_888x873.png 424w, https://substackcdn.com/image/fetch/$s_!cIpN!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3914a985-e6fc-452d-8332-24e2c7f38ea5_888x873.png 848w, https://substackcdn.com/image/fetch/$s_!cIpN!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3914a985-e6fc-452d-8332-24e2c7f38ea5_888x873.png 1272w, https://substackcdn.com/image/fetch/$s_!cIpN!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3914a985-e6fc-452d-8332-24e2c7f38ea5_888x873.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!cIpN!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3914a985-e6fc-452d-8332-24e2c7f38ea5_888x873.png" width="888" height="873" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/3914a985-e6fc-452d-8332-24e2c7f38ea5_888x873.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:873,&quot;width&quot;:888,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:434108,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!cIpN!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3914a985-e6fc-452d-8332-24e2c7f38ea5_888x873.png 424w, https://substackcdn.com/image/fetch/$s_!cIpN!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3914a985-e6fc-452d-8332-24e2c7f38ea5_888x873.png 848w, https://substackcdn.com/image/fetch/$s_!cIpN!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3914a985-e6fc-452d-8332-24e2c7f38ea5_888x873.png 1272w, https://substackcdn.com/image/fetch/$s_!cIpN!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3914a985-e6fc-452d-8332-24e2c7f38ea5_888x873.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><a href="https://arxiv.org/abs/2502.04363">https://arxiv.org/abs/2502.04363</a></p><p>The challenge lies in running demanding text-to-video generation models on resource-limited mobile devices. Current diffusion-based video generation is computationally and memory intensive, making it inaccessible on smartphones.</p><p>This paper introduces On-device Sora. It is a pioneering solution to enable diffusion-based text-to-video generation directly on mobile devices. On-device Sora employs three novel techniques to overcome resource constraints, maintaining video quality comparable to server-grade performance.</p><p>-----</p><p>&#128204; Linear Proportional Leap effectively reduces diffusion steps by half. This method smartly exploits Rectified Flow's linear trajectory. It significantly accelerates video generation on devices without retraining.</p><p>&#128204; Temporal Dimension Token Merging cleverly merges tokens along the time dimension. This reduces computational load in attention layers by a quarter. It maintains video quality while boosting efficiency.</p><p>&#128204; Concurrent Inference with Dynamic Loading overcomes memory constraints on devices. It allows for efficient execution of large models by concurrent processing and dynamic block management.</p><p>----------</p><p>Methods Explored in this Paper &#128295;:</p><p>&#8594; Linear Proportional Leap (LPL) is introduced to reduce the number of denoising steps. LPL leverages the linear trajectory of Rectified Flow to make direct "leaps" in the denoising process. This avoids performing all originally required denoising steps.</p><p>&#8594; Temporal Dimension Token Merging (TDTM) is proposed to minimize token processing computation. TDTM merges consecutive video frame tokens in the temporal dimension within the attention layers of the Spatial Temporal Diffusion Transformer (STDiT). This reduces the number of tokens and computational load.</p><p>&#8594; Concurrent Inference with Dynamic Loading (CI-DL) addresses memory limitations. CI-DL partitions large models into smaller blocks. It then dynamically loads these blocks into memory for concurrent model inference and block loading. This optimizes memory use and speeds up processing.</p><p>-----</p><p>Key Insights &#128161;:</p><p>&#8594; On-device Sora is the first framework to enable high-quality diffusion-based text-to-video generation on smartphones.</p><p>&#8594; The proposed methods, LPL, TDTM, and CI-DL, significantly improve efficiency without compromising video quality.</p><p>&#8594; On-device video generation enhances user privacy, reduces cloud dependency, and lowers costs.</p><p>-----</p><p>Results &#128202;:</p><p>&#8594; Achieves up to 1.94&#215; speedup using Linear Proportional Leap (LPL) while maintaining comparable video quality.</p><p>&#8594; Temporal Dimension Token Merging (TDTM) provides up to 1.27&#215; speedup with stable video quality metrics.</p><p>&#8594; Concurrent Inference with Dynamic Loading (CI-DL) reduces STDiT inference latency by approximately 25%, from 1000 to 750 seconds for 30 denoising steps.</p>]]></content:encoded></item><item><title><![CDATA["Ola: Pushing the Frontiers of Omni-Modal Language Model with Progressive Modality Alignment"]]></title><description><![CDATA[Below podcast on this paper is generated with Google's Illuminate.]]></description><link>https://www.rohan-paul.com/p/ola-pushing-the-frontiers-of-omni-836</link><guid isPermaLink="false">https://www.rohan-paul.com/p/ola-pushing-the-frontiers-of-omni-836</guid><dc:creator><![CDATA[Rohan Paul]]></dc:creator><pubDate>Sat, 15 Feb 2025 14:18:24 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!MvV2!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9c2bd4a8-92e9-43a9-a006-d8fb24d4c334_808x867.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!MvV2!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9c2bd4a8-92e9-43a9-a006-d8fb24d4c334_808x867.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!MvV2!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9c2bd4a8-92e9-43a9-a006-d8fb24d4c334_808x867.png 424w, https://substackcdn.com/image/fetch/$s_!MvV2!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9c2bd4a8-92e9-43a9-a006-d8fb24d4c334_808x867.png 848w, https://substackcdn.com/image/fetch/$s_!MvV2!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9c2bd4a8-92e9-43a9-a006-d8fb24d4c334_808x867.png 1272w, https://substackcdn.com/image/fetch/$s_!MvV2!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9c2bd4a8-92e9-43a9-a006-d8fb24d4c334_808x867.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!MvV2!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9c2bd4a8-92e9-43a9-a006-d8fb24d4c334_808x867.png" width="808" height="867" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/9c2bd4a8-92e9-43a9-a006-d8fb24d4c334_808x867.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:867,&quot;width&quot;:808,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:285527,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!MvV2!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9c2bd4a8-92e9-43a9-a006-d8fb24d4c334_808x867.png 424w, https://substackcdn.com/image/fetch/$s_!MvV2!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9c2bd4a8-92e9-43a9-a006-d8fb24d4c334_808x867.png 848w, https://substackcdn.com/image/fetch/$s_!MvV2!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9c2bd4a8-92e9-43a9-a006-d8fb24d4c334_808x867.png 1272w, https://substackcdn.com/image/fetch/$s_!MvV2!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9c2bd4a8-92e9-43a9-a006-d8fb24d4c334_808x867.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><a href="https://arxiv.org/abs/2502.04328">https://arxiv.org/abs/2502.04328</a></p><p>The challenge in creating omni-modal models is their performance gap compared to specialized single modality models. Existing omni-modal models also lack balanced performance and efficient training.</p><p>This paper introduces Ola, an Omni-modal Language Model, using progressive modality alignment to address these issues. Ola achieves competitive performance across image, video, and audio tasks.</p><p>-----</p><p>&#128204; Progressive modality alignment in Ola effectively manages multi-modal interference. Training starts with text-image pairs. Then it progressively integrates audio and video. This staged approach avoids catastrophic forgetting and improves overall performance.</p><p>&#128204; Ola's architecture uses modality-specific encoders with a shared LLM decoder. This design choice allows efficient feature extraction from diverse inputs. The Local-Global Attention Pooling further optimizes visual token processing for better efficiency.</p><p>&#128204; Generating cross-modal video-audio data is a key innovation. This data bridges the gap between vision and audio. It enhances the model's ability to understand real-world scenarios where modalities are naturally correlated.</p><p>----------</p><p>Methods Explored in this Paper &#128295;:</p><p>&#8594; Ola employs a progressive modality alignment strategy.</p><p>&#8594; It begins training with image and text data. This establishes core vision-language knowledge.</p><p>&#8594; Then, speech data is added. Speech bridges language and audio understanding.</p><p>&#8594; Finally, video data is incorporated. Video connects all modalities: vision, audio, and language.</p><p>&#8594; Ola's architecture includes modality-specific encoders. These handle text, image, video, and audio inputs.</p><p>&#8594; A Local-Global Attention Pooling layer fuses visual inputs. This reduces visual token length efficiently.</p><p>&#8594; Dual audio encoders are used. Whisper for speech and BEATs for music are used. This allows for richer audio understanding.</p><p>&#8594; Sentence-wise streaming decoding is implemented for real-time text and speech generation.</p><p>-----</p><p>Key Insights &#128161;:</p><p>&#8594; Progressive modality alignment improves omni-modal learning. It breaks down complex training into manageable steps.</p><p>&#8594; This strategy maintains a small cross-modal alignment data size. It leverages existing vision-language model advancements.</p><p>&#8594; Video data acts as a crucial bridge between audio and vision. It provides comprehensive multi-modal information.</p><p>&#8594; Cross-modal video-audio data generation enhances performance. This data captures relationships between video and audio.</p><p>-----</p><p>Results &#128202;:</p><p>&#8594; Ola achieves 84.3% accuracy on MMBench-1.1.</p><p>&#8594; Ola achieves 70.8% accuracy on MMStar.</p><p>&#8594; Ola achieves 57.0% accuracy on MMMU.</p><p>&#8594; Ola achieves 68.4% accuracy on VideoMME.</p><p>&#8594; Ola achieves 3.1 mean WER on LibriSpeech.</p><p>&#8594; Ola achieves 6.41 GPT-eval score on AIR-Bench.</p>]]></content:encoded></item><item><title><![CDATA["Ola: Pushing the Frontiers of Omni-Modal Language Model with Progressive Modality Alignment"]]></title><description><![CDATA[Below podcast on this paper is generated with Google's Illuminate.]]></description><link>https://www.rohan-paul.com/p/ola-pushing-the-frontiers-of-omni</link><guid isPermaLink="false">https://www.rohan-paul.com/p/ola-pushing-the-frontiers-of-omni</guid><dc:creator><![CDATA[Rohan Paul]]></dc:creator><pubDate>Sat, 15 Feb 2025 13:53:38 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!xMFH!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb909cc01-c7f3-4003-a073-74e4018e6813_808x867.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!xMFH!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb909cc01-c7f3-4003-a073-74e4018e6813_808x867.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!xMFH!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb909cc01-c7f3-4003-a073-74e4018e6813_808x867.png 424w, https://substackcdn.com/image/fetch/$s_!xMFH!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb909cc01-c7f3-4003-a073-74e4018e6813_808x867.png 848w, https://substackcdn.com/image/fetch/$s_!xMFH!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb909cc01-c7f3-4003-a073-74e4018e6813_808x867.png 1272w, https://substackcdn.com/image/fetch/$s_!xMFH!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb909cc01-c7f3-4003-a073-74e4018e6813_808x867.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!xMFH!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb909cc01-c7f3-4003-a073-74e4018e6813_808x867.png" width="808" height="867" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b909cc01-c7f3-4003-a073-74e4018e6813_808x867.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:867,&quot;width&quot;:808,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:285527,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!xMFH!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb909cc01-c7f3-4003-a073-74e4018e6813_808x867.png 424w, https://substackcdn.com/image/fetch/$s_!xMFH!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb909cc01-c7f3-4003-a073-74e4018e6813_808x867.png 848w, https://substackcdn.com/image/fetch/$s_!xMFH!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb909cc01-c7f3-4003-a073-74e4018e6813_808x867.png 1272w, https://substackcdn.com/image/fetch/$s_!xMFH!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb909cc01-c7f3-4003-a073-74e4018e6813_808x867.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><a href="https://arxiv.org/abs/2502.04328">https://arxiv.org/abs/2502.04328</a></p><p>The challenge in creating omni-modal models is their performance gap compared to specialized single modality models. Existing omni-modal models also lack balanced performance and efficient training.</p><p>This paper introduces Ola, an Omni-modal Language Model, using progressive modality alignment to address these issues. Ola achieves competitive performance across image, video, and audio tasks.</p><p>-----</p><p>&#128204; Progressive modality alignment in Ola effectively manages multi-modal interference. Training starts with text-image pairs. Then it progressively integrates audio and video. This staged approach avoids catastrophic forgetting and improves overall performance.</p><p>&#128204; Ola's architecture uses modality-specific encoders with a shared LLM decoder. This design choice allows efficient feature extraction from diverse inputs. The Local-Global Attention Pooling further optimizes visual token processing for better efficiency.</p><p>&#128204; Generating cross-modal video-audio data is a key innovation. This data bridges the gap between vision and audio. It enhances the model's ability to understand real-world scenarios where modalities are naturally correlated.</p><p>----------</p><p>Methods Explored in this Paper &#128295;:</p><p>&#8594; Ola employs a progressive modality alignment strategy.</p><p>&#8594; It begins training with image and text data. This establishes core vision-language knowledge.</p><p>&#8594; Then, speech data is added. Speech bridges language and audio understanding.</p><p>&#8594; Finally, video data is incorporated. Video connects all modalities: vision, audio, and language.</p><p>&#8594; Ola's architecture includes modality-specific encoders. These handle text, image, video, and audio inputs.</p><p>&#8594; A Local-Global Attention Pooling layer fuses visual inputs. This reduces visual token length efficiently.</p><p>&#8594; Dual audio encoders are used. Whisper for speech and BEATs for music are used. This allows for richer audio understanding.</p><p>&#8594; Sentence-wise streaming decoding is implemented for real-time text and speech generation.</p><p>-----</p><p>Key Insights &#128161;:</p><p>&#8594; Progressive modality alignment improves omni-modal learning. It breaks down complex training into manageable steps.</p><p>&#8594; This strategy maintains a small cross-modal alignment data size. It leverages existing vision-language model advancements.</p><p>&#8594; Video data acts as a crucial bridge between audio and vision. It provides comprehensive multi-modal information.</p><p>&#8594; Cross-modal video-audio data generation enhances performance. This data captures relationships between video and audio.</p><p>-----</p><p>Results &#128202;:</p><p>&#8594; Ola achieves 84.3% accuracy on MMBench-1.1.</p><p>&#8594; Ola achieves 70.8% accuracy on MMStar.</p><p>&#8594; Ola achieves 57.0% accuracy on MMMU.</p><p>&#8594; Ola achieves 68.4% accuracy on VideoMME.</p><p>&#8594; Ola achieves 3.1 mean WER on LibriSpeech.</p><p>&#8594; Ola achieves 6.41 GPT-eval score on AIR-Bench.</p>]]></content:encoded></item><item><title><![CDATA["No Task Left Behind: Isotropic Model Merging with Common and Task-Specific Subspaces"]]></title><description><![CDATA[Below podcast on this paper is generated with Google's Illuminate.]]></description><link>https://www.rohan-paul.com/p/no-task-left-behind-isotropic-model</link><guid isPermaLink="false">https://www.rohan-paul.com/p/no-task-left-behind-isotropic-model</guid><dc:creator><![CDATA[Rohan Paul]]></dc:creator><pubDate>Sat, 15 Feb 2025 13:49:31 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!1L4N!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F082b6fa2-141c-4a3f-a6c9-1c340e09cf2e_1083x873.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!1L4N!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F082b6fa2-141c-4a3f-a6c9-1c340e09cf2e_1083x873.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!1L4N!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F082b6fa2-141c-4a3f-a6c9-1c340e09cf2e_1083x873.png 424w, https://substackcdn.com/image/fetch/$s_!1L4N!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F082b6fa2-141c-4a3f-a6c9-1c340e09cf2e_1083x873.png 848w, https://substackcdn.com/image/fetch/$s_!1L4N!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F082b6fa2-141c-4a3f-a6c9-1c340e09cf2e_1083x873.png 1272w, https://substackcdn.com/image/fetch/$s_!1L4N!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F082b6fa2-141c-4a3f-a6c9-1c340e09cf2e_1083x873.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!1L4N!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F082b6fa2-141c-4a3f-a6c9-1c340e09cf2e_1083x873.png" width="1083" height="873" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/082b6fa2-141c-4a3f-a6c9-1c340e09cf2e_1083x873.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:873,&quot;width&quot;:1083,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:360102,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!1L4N!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F082b6fa2-141c-4a3f-a6c9-1c340e09cf2e_1083x873.png 424w, https://substackcdn.com/image/fetch/$s_!1L4N!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F082b6fa2-141c-4a3f-a6c9-1c340e09cf2e_1083x873.png 848w, https://substackcdn.com/image/fetch/$s_!1L4N!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F082b6fa2-141c-4a3f-a6c9-1c340e09cf2e_1083x873.png 1272w, https://substackcdn.com/image/fetch/$s_!1L4N!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F082b6fa2-141c-4a3f-a6c9-1c340e09cf2e_1083x873.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><a href="https://arxiv.org/abs/2502.04959">https://arxiv.org/abs/2502.04959</a></p><p>The problem is that merging multiple task-specific models into one multi-task model still results in a performance drop compared to individual task-specific models. This paper introduces a new merging method to reduce this performance gap by analyzing the properties of task matrices.</p><p>This paper proposes "Isotropic Merging". It flattens the singular value spectrum of merged task matrices to improve alignment between task-specific and merged subspaces. This approach enhances performance without additional training.</p><p>-----</p><p>Okay, here are my technical perspectives on the paper's solution:</p><p>&#128204; Isotropic Merging cleverly uses Singular Value Decomposition to flatten the singular value spectrum. This balances task representation in merged models, boosting multi-task performance without extra training.</p><p>&#128204; The paper shifts focus from cosine similarity to subspace alignment for effective merging. By aligning task and merged subspaces, Isotropic Merging achieves superior performance compared to Task Arithmetic.</p><p>&#128204; Iso-CTS method practically combines common and task-specific subspaces. This simple yet effective approach reaches state-of-the-art model merging results across diverse tasks and model scales.</p><p>----------</p><p>Methods Explored in this Paper &#128295;:</p><p>&#8594; This paper explores model merging through the lens of Singular Value Decomposition of task matrices. Task matrix is the weight update matrix applied to a pre-trained model for a specific task.</p><p>&#8594; Introduces Subspace Alignment Ratio metric. It quantifies the similarity between subspaces of task-specific matrices and merged matrices.</p><p>&#8594; Isotropic Merging in Common Subspace (Iso-C) is proposed. Iso-C flattens the singular value spectrum of the merged task matrix to make it uniform. This is done by scaling all singular directions to the average singular value.</p><p>&#8594; Isotropic Merging in Common and Task-Specific Subspaces (Iso-CTS) is also proposed. Iso-CTS enhances Iso-C by incorporating task-specific directions. It retains top singular vectors from a common subspace and adds task-specific directions orthogonal to the common subspace.</p><p>-----</p><p>Key Insights &#128161;:</p><p>&#8594; Alignment between singular components of task-specific and merged matrices strongly correlates with performance improvement.</p><p>&#8594; Flattening the singular value spectrum of merged matrices improves subspace alignment. This leads to better multi-task performance.</p><p>&#8594; Incorporating task-specific subspaces along with a common subspace further enhances merging performance, especially with a larger number of tasks.</p><p>-----</p><p>Results &#128202;:</p><p>&#8594; Iso-CTS achieves state-of-the-art performance across various tasks and model scales (ViT-B/32, ViT-B/16, ViT-L/14).</p><p>&#8594; Iso-CTS outperforms Task Arithmetic by a large margin, especially when merging 14 and 20 tasks, with up to 2.8% absolute accuracy improvement.</p><p>&#8594; Iso-C and Iso-CTS are more robust to the choice of scaling factor &#945; compared to Task Arithmetic.</p>]]></content:encoded></item><item><title><![CDATA["MotionLab: Unified Human Motion Generation and Editing via the Motion-Condition-Motion Paradigm"]]></title><description><![CDATA[Below podcast on this paper is generated with Google's Illuminate.]]></description><link>https://www.rohan-paul.com/p/motionlab-unified-human-motion-generation</link><guid isPermaLink="false">https://www.rohan-paul.com/p/motionlab-unified-human-motion-generation</guid><dc:creator><![CDATA[Rohan Paul]]></dc:creator><pubDate>Sat, 15 Feb 2025 13:46:31 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!R85x!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff8083393-f443-42b4-ac3c-f9724593ec82_890x810.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!R85x!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff8083393-f443-42b4-ac3c-f9724593ec82_890x810.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!R85x!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff8083393-f443-42b4-ac3c-f9724593ec82_890x810.png 424w, https://substackcdn.com/image/fetch/$s_!R85x!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff8083393-f443-42b4-ac3c-f9724593ec82_890x810.png 848w, https://substackcdn.com/image/fetch/$s_!R85x!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff8083393-f443-42b4-ac3c-f9724593ec82_890x810.png 1272w, https://substackcdn.com/image/fetch/$s_!R85x!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff8083393-f443-42b4-ac3c-f9724593ec82_890x810.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!R85x!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff8083393-f443-42b4-ac3c-f9724593ec82_890x810.png" width="890" height="810" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f8083393-f443-42b4-ac3c-f9724593ec82_890x810.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:810,&quot;width&quot;:890,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:538425,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!R85x!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff8083393-f443-42b4-ac3c-f9724593ec82_890x810.png 424w, https://substackcdn.com/image/fetch/$s_!R85x!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff8083393-f443-42b4-ac3c-f9724593ec82_890x810.png 848w, https://substackcdn.com/image/fetch/$s_!R85x!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff8083393-f443-42b4-ac3c-f9724593ec82_890x810.png 1272w, https://substackcdn.com/image/fetch/$s_!R85x!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff8083393-f443-42b4-ac3c-f9724593ec82_890x810.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><a href="https://arxiv.org/abs/2502.02358">https://arxiv.org/abs/2502.02358</a></p><p>The paper addresses the challenge of isolated solutions for human motion generation and editing tasks. Current methods lack versatility, fine-grained control and knowledge sharing across tasks.</p><p>This paper introduces MotionLab, a unified framework based on the Motion-Condition-Motion paradigm. MotionLab uses rectified flows and a MotionFlow Transformer to map source motion to target motion, guided by conditions. It incorporates Aligned Rotational Position Encoding and Task Instruction Modulation for effective multi-task learning via Motion Curriculum Learning.</p><p>-----</p><p>&#128204; Motion-Condition-Motion paradigm offers a simple yet effective abstraction. It unifies motion generation and editing. Rectified flows enable efficient mapping between source and target motions.</p><p>&#128204; MotionFlow Transformer with Joint Attention and Condition Path is key. It facilitates multi-modal interaction. Adaptive Layer Normalization enhances conditional control without task-specific modules.</p><p>&#128204; Aligned Rotational Position Encoding addresses temporal misalignment. It is crucial for time-sensitive motion tasks. Motion Curriculum Learning enables effective multi-task training.</p><p>----------</p><p>Methods Explored in this Paper &#128295;:</p><p>&#8594; MotionLab uses the Motion-Condition-Motion paradigm. This paradigm unifies motion generation and editing tasks using source motion, condition, and target motion concepts.</p><p>&#8594; MotionLab framework is built around the MotionFlow Transformer (MFT). MFT leverages rectified flows to learn the mapping from source motion to target motion based on conditions.</p><p>&#8594; MFT includes Joint Attention to enable interaction between tokens from different modalities. A Condition Path differentiates modalities and extracts representations. Aligned Rotational Position Encoding (ROPE) ensures time synchronization.</p><p>&#8594; Task Instruction Modulation is used to differentiate tasks by adding a task-specific instruction embedding into the MFT.</p><p>&#8594; Motion Curriculum Learning is a training strategy. It organizes tasks by difficulty for effective multi-task learning and knowledge sharing. Training is divided into masked pre-training and supervised fine-tuning stages.</p><p>-----</p><p>Key Insights &#128161;:</p><p>&#8594; The Motion-Condition-Motion paradigm effectively unifies diverse human motion generation and editing tasks.</p><p>&#8594; MotionLab framework, with its MotionFlow Transformer and Motion Curriculum Learning, achieves versatility and strong performance across various motion tasks.</p><p>&#8594; Aligned ROPE is crucial for maintaining temporal synchronization between source and target motions, improving performance in time-sensitive tasks.</p><p>&#8594; Task Instruction Modulation and Motion Curriculum Learning are essential for effective multi-task learning and knowledge sharing across different motion tasks.</p><p>-----</p><p>Results &#128202;:</p><p>&#8594; MotionLab achieves a FID score of 0.223 in text-based motion generation on the HumanML3D dataset.</p><p>&#8594; In trajectory-based motion generation on HumanML3D, MotionLab achieves an average error of 0.0334 when controlling all joints.</p><p>&#8594; For text-based motion editing on the MotionFix dataset, MotionLab attains a retrieval rate (R@1) of 56.34%.</p><p>&#8594; In trajectory-based motion editing on MotionFix, MotionLab reaches a retrieval rate (R@1) of 72.65%.</p><p>&#8594; MotionLab achieves a Style Recognition Accuracy (SRA) of 64.97% in motion style transfer.</p>]]></content:encoded></item><item><title><![CDATA["MotionCanvas: Cinematic Shot Design with Controllable Image-to-Video Generation"]]></title><description><![CDATA[Below podcast on this paper is generated with Google's Illuminate.]]></description><link>https://www.rohan-paul.com/p/motioncanvas-cinematic-shot-design</link><guid isPermaLink="false">https://www.rohan-paul.com/p/motioncanvas-cinematic-shot-design</guid><dc:creator><![CDATA[Rohan Paul]]></dc:creator><pubDate>Sat, 15 Feb 2025 13:42:39 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!2263!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f155073-2c1c-4de0-9237-3cf3ea238749_763x854.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!2263!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f155073-2c1c-4de0-9237-3cf3ea238749_763x854.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!2263!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f155073-2c1c-4de0-9237-3cf3ea238749_763x854.png 424w, https://substackcdn.com/image/fetch/$s_!2263!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f155073-2c1c-4de0-9237-3cf3ea238749_763x854.png 848w, https://substackcdn.com/image/fetch/$s_!2263!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f155073-2c1c-4de0-9237-3cf3ea238749_763x854.png 1272w, https://substackcdn.com/image/fetch/$s_!2263!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f155073-2c1c-4de0-9237-3cf3ea238749_763x854.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!2263!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f155073-2c1c-4de0-9237-3cf3ea238749_763x854.png" width="763" height="854" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/1f155073-2c1c-4de0-9237-3cf3ea238749_763x854.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:854,&quot;width&quot;:763,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:474280,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!2263!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f155073-2c1c-4de0-9237-3cf3ea238749_763x854.png 424w, https://substackcdn.com/image/fetch/$s_!2263!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f155073-2c1c-4de0-9237-3cf3ea238749_763x854.png 848w, https://substackcdn.com/image/fetch/$s_!2263!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f155073-2c1c-4de0-9237-3cf3ea238749_763x854.png 1272w, https://substackcdn.com/image/fetch/$s_!2263!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f155073-2c1c-4de0-9237-3cf3ea238749_763x854.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><a href="https://arxiv.org/abs/2502.04299">https://arxiv.org/abs/2502.04299</a></p><p>The paper addresses the problem of limited user control in image-to-video generation, particularly for cinematic shot design involving both camera and object motions. Current methods lack intuitive and precise control over these intertwined movements, hindering creative expression.</p><p>This paper proposes MotionCanvas, a system that enables cinematic shot design by integrating user-driven controls for camera and object motions into image-to-video generation models. It uses a Motion Signal Translation module to bridge user intent and model conditioning.</p><p>-----</p><p>&#128204; MotionCanvas introduces a crucial Motion Signal Translation module. This module smartly converts user-friendly 3D motion intents into 2D screen-space signals. This bridging is key for effective video diffusion model control.</p><p>&#128204; The strength of MotionCanvas lies in achieving 3D-aware cinematic control using only 2D training data. This bypasses the need for costly 3D annotations. It leverages depth estimation and geometric transformations for scene understanding.</p><p>&#128204; MotionCanvas offers immediate practical value. It empowers users to design complex cinematic shots with combined camera and object motions. This enhanced control improves creative workflows in video content creation tasks.</p><p>----------</p><p>Methods Explored in this Paper &#128295;:</p><p>&#8594; MotionCanvas uses a Motion Design Module to capture user intentions for camera motion, object global motion, and object local motion. Users can define camera paths with 3D poses or base motion patterns. Scene-anchored bounding boxes control object global motion. Point trajectories control object local motion. Timing control is also integrated.</p><p>&#8594; The Motion Signal Translation module converts scene-space motion designs into screen-space motion signals. Camera movement is represented by 2D point tracks derived from 3D camera paths using depth estimation. Scene-aware object motion uses bounding box trajectories converted to screen space, considering camera motion. Object local motion uses point trajectories transformed to screen space, accounting for both camera and global object motion.</p><p>&#8594; A motion-conditioned video generation model, based on Diffusion Transformer (DiT), is fine-tuned with these screen-space motion signals. Point trajectories are encoded using Discrete Cosine Transform (DCT) coefficients. Bounding box sequences are encoded as color-coded masks. Auto-regressive generation is used for variable-length videos.</p><p>-----</p><p>Key Insights &#128161;:</p><p>&#8594; Cinematic shot design requires simultaneous and intuitive control over camera and object motions in image-to-video generation.</p><p>&#8594; MotionCanvas effectively translates user-defined scene-space motion intent into screen-space control signals for video diffusion models.</p><p>&#8594; Representing camera motion with point tracking and object motion with scene-anchored bounding boxes allows for 3D-aware control without explicit 3D training data.</p><p>&#8594; DCT coefficients and color-coded masks provide efficient and effective ways to condition video diffusion models on motion signals.</p><p>-----</p><p>Results &#128202;:</p><p>&#8594; MotionCanvas achieves lower Rotation Error (0.6334) and Translation Error (0.2188) in camera motion control compared to MotionCtrl and CameraCtrl.</p><p>&#8594; MotionCanvas reduces Frechet Video Distance (FVD) to 34.09 and Frechet Inception Distance (FID) to 7.60, indicating higher video quality.</p><p>&#8594; User studies show MotionCanvas is preferred for motion adherence (75.24%), motion quality (79.05%), and frame fidelity (77.14%) over DragAnything and MOFA-Video.</p>]]></content:encoded></item></channel></rss>