<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[Ankur’s Newsletter]]></title><description><![CDATA[Covering major trends in artificial intelligence and startups.]]></description><link>https://www.ankursnewsletter.com</link><image><url>https://substackcdn.com/image/fetch/$s_!h-fI!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2793f3d2-b75e-4405-abed-a419427aca14_900x900.png</url><title>Ankur’s Newsletter</title><link>https://www.ankursnewsletter.com</link></image><generator>Substack</generator><lastBuildDate>Thu, 30 Apr 2026 04:25:45 GMT</lastBuildDate><atom:link href="https://www.ankursnewsletter.com/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Ankur A. Patel]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[ankurpatel@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[ankurpatel@substack.com]]></itunes:email><itunes:name><![CDATA[Ankur A. Patel]]></itunes:name></itunes:owner><itunes:author><![CDATA[Ankur A. Patel]]></itunes:author><googleplay:owner><![CDATA[ankurpatel@substack.com]]></googleplay:owner><googleplay:email><![CDATA[ankurpatel@substack.com]]></googleplay:email><googleplay:author><![CDATA[Ankur A. Patel]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[GPT-5 in Production: What Early Enterprise Integrations Reveal]]></title><description><![CDATA[GPT-5's enterprise rollouts boost productivity up to 40%, but costs, latency and governance risks demand tight token limits, guardrails and hybrid designs.]]></description><link>https://www.ankursnewsletter.com/p/gpt-5-in-production-what-early-enterprise</link><guid isPermaLink="false">https://www.ankursnewsletter.com/p/gpt-5-in-production-what-early-enterprise</guid><dc:creator><![CDATA[Ankur A. Patel]]></dc:creator><pubDate>Thu, 21 Aug 2025 21:28:00 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!4WwW!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1550a706-61ba-4afb-8e5e-715081088755_1200x600.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!4WwW!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1550a706-61ba-4afb-8e5e-715081088755_1200x600.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!4WwW!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1550a706-61ba-4afb-8e5e-715081088755_1200x600.png 424w, https://substackcdn.com/image/fetch/$s_!4WwW!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1550a706-61ba-4afb-8e5e-715081088755_1200x600.png 848w, https://substackcdn.com/image/fetch/$s_!4WwW!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1550a706-61ba-4afb-8e5e-715081088755_1200x600.png 1272w, https://substackcdn.com/image/fetch/$s_!4WwW!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1550a706-61ba-4afb-8e5e-715081088755_1200x600.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!4WwW!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1550a706-61ba-4afb-8e5e-715081088755_1200x600.png" width="1200" height="600" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/1550a706-61ba-4afb-8e5e-715081088755_1200x600.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:600,&quot;width&quot;:1200,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:382110,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.ankursnewsletter.com/i/171599148?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1550a706-61ba-4afb-8e5e-715081088755_1200x600.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!4WwW!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1550a706-61ba-4afb-8e5e-715081088755_1200x600.png 424w, https://substackcdn.com/image/fetch/$s_!4WwW!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1550a706-61ba-4afb-8e5e-715081088755_1200x600.png 848w, https://substackcdn.com/image/fetch/$s_!4WwW!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1550a706-61ba-4afb-8e5e-715081088755_1200x600.png 1272w, https://substackcdn.com/image/fetch/$s_!4WwW!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1550a706-61ba-4afb-8e5e-715081088755_1200x600.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.ankursnewsletter.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Subscribe today to join our community of 2030 AI enthusiasts and business leaders for more insider takes and practical advice on implementing AI in business.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h1>Key Takeaways</h1><ul><li><p>GPT-5&#8217;s 6-trillion-parameter MoE core delivers <strong>stronger multi-step reasoning</strong> and 256 k context windows, but latency and cost rise steeply on prompts beyond ~80 k tokens.</p></li><li><p>Early enterprise pilots (Microsoft Copilot, Apple Intelligence, Amgen, BBVA) show 20&#8211;40% <strong>productivity gains only when strict token limits</strong>, vector-RAG grounding, and deterministic tool calls are in place.</p></li><li><p>The <strong>real-time router and selective fine-tuning lower blended costs by off-loading simple tasks to smaller models</strong> and updating just 15% of weights, yet careless usage can erase those savings.</p></li><li><p>Guardrail embeddings cut jailbreaks below 1%, but over-blocking and the 2024 knowledge cutoff mean <strong>human oversight, policy layers, and zero-trust design remain mandatory.</strong></p></li><li><p>CTOs should treat GPT-5 as a <strong>high-governance, high-value engine</strong>&#8212;benchmark against small LLMs, automate evals and guardrails, and prepare hybrid edge-cloud paths to balance privacy, latency, and spend.</p></li></ul><div><hr></div><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!R9ua!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!R9ua!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!R9ua!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png" width="1456" height="246" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/bea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:246,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:273651,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:&quot;&quot;,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!R9ua!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1456w" sizes="100vw" loading="lazy" fetchpriority="high"></picture><div></div></div></a></figure></div><p>In 2023, I started Multimodal, a Generative AI company that helps organizations automate complex, knowledge-based workflows using AI Agents. Check it out <a href="https://www.multimodal.dev/">here</a>.</p><div><hr></div><p><a href="https://openai.com/index/introducing-gpt-5/">OpenAI positions </a><em><a href="https://openai.com/index/introducing-gpt-5/">GPT-5</a></em> as its first truly <strong>enterprise-grade</strong> LLM, yet the upgrade is mixed. The 6-trillion-parameter core on a compressed MoE router keeps latency close to GPT-4o, but only when prompts stay under 80 k tokens; beyond that I see two-to-three-second spikes. The 256 k context accepts <em>documents</em>, <em>images</em>, <em>code</em> traces, and sensor <em>data</em>, returning a strict <strong>JSON</strong> schema that slips into any REST <em>API</em>.</p><p>Synthetic execution-trace pre-training lifts <em>multi-step reasoning</em> accuracy by roughly 35 percent in OpenAI&#8217;s public &#8220;AgentBench&#8221; scores, but real-world variance is high. In controlled tests the new <em>Guardrail Embeddings</em> cut jailbreaks to below one percent&#8212;helpful, though they occasionally block harmless financial-model prompts. <strong>The </strong><em><strong>real-time router</strong></em><strong> diverts simple </strong><em><strong>tasks</strong></em><strong> to a smaller helper model, saving about 0.3 &#162; per 1 k tokens, but that gain disappears if your workload leans on long-form analysis.</strong></p><p><strong>My view</strong>: GPT-5 solves predictable <em>function calling</em> better than <em>other models</em>, yet its <em>cost</em> profile and occasional throttling mean it is not a default drop-in replacement. Teams should benchmark against task-specific SLMs before committing a budget.</p><h2>Deployment Landscape (August 2025)</h2><h3>Microsoft Copilot (M365 E5 preview)</h3><p>Word, Teams, and Power BI now default to <strong>OpenAI GPT 5</strong> for long prompts. Azure AI &#8220;Orion&#8221; clusters report a <strong>40% GPU-cycle efficiency gain</strong> from speculative decoding, holding latency near 800 ms for 12 k-token jobs. In my tests the <em>real time router</em> still falls back to smaller <em>models</em> when <em>tasks</em> need fewer than 1 k <em>output tokens</em>, saving about 18% on <em>cost</em>.</p><h3>Apple Intelligence (iOS 19 beta)</h3><p>Apple&#8217;s Ajax-M4 handles local prompts, but any input above 4 k tokens ships to GPT-5 over a Private Relay tunnel. No Apple ID metadata is stored, although the hybrid path adds roughly 250 ms. I see stronger multi step reasoning than other models, yet advanced image-code blends remain limited by Apple&#8217;s privacy wrapper.</p><h3>Amgen R&amp;D Assistant</h3><p> A secure enclave on AWS GB200 chips hosts GPT-5 fine-tuned with 200 k ELN <em>documents</em>. Through REST <em>API</em> calls the agent cuts protocol-draft turnaround by 31%. Compliance teams flag rare false positives in chemical nomenclature, likely tied to the 2024 <em>knowledge cut off</em>.</p><p><em>BBVA Finance Copilot<br></em> Spanish&#8211;English prompts route through SAP HANA adapters. GPT-5&#8217;s structured <em>answers</em> update exposure tables in real time. The bank caps each query at 8 k tokens to control <em>spend</em>.</p><p>GPT-5 is the <strong>right model</strong> for these high-governance settings, but only when <em>developers</em> embed strict token and privacy controls upfront.</p><h2>Early Wins</h2><p>Early-access deployments show that <strong>open ai gpt 5</strong> can deliver material efficiency gains, though most benefits hinge on tight vertical grounding rather than raw <em>advanced reasoning</em> headroom.</p><ul><li><p>Amgen: Regulatory dossier drafting now closes in 7 days, down from 11, a <strong>31 percent</strong> acceleration. The assistant ingests 120 k&#8208;token clinical <em>documents</em>, taps ELN metadata through a REST <em>API</em>, then autowrites the boiler-plate sections. The <em>model</em>&#8217;s longer context means fewer fragment transitions, so editors spend less <em>effort</em> stitching <em>excel</em> tables or <em>code</em> snippets back together.<br></p></li><li><p>Microsoft Copilot: Power BI teams report a <strong>22 percent</strong> reduction in dashboard build time. GPT-5 generates DAX formulas on the first <em>prompt</em> more often than <em>other models</em> thanks to better <em>multi step reasoning</em>. When the <em>real time router</em> off-loads small <em>tasks</em> to a lightweight sibling, <em>cost</em> per interactive session falls by about 18 percent.<br></p></li><li><p>Apple Intelligence beta: Test <em>users</em> see Siri follow-through double on chained commands (&#8220;text my partner, book rideshare, start playlist&#8221;). Hybrid on-device inference handles the simple &#8220;text&#8221; step, then cloud GPT-5 processes the multi-turn <em>plan</em>. Fewer missed intents translate into measurable <em>productivity</em> gains for everyday routines.</p></li></ul><p>I view these results as examples of the right<em> model, right glue</em>. The vector-DB layer supplies domain <em>data</em>, GPT-5 adds <em>reasoning capabilities</em>, and deterministic tool calls produce outputs the back-end can trust. The boost is not magic thinking or some vague IQ spike; it is disciplined system design.</p><h2>Pain Points &amp; Bottlenecks</h2><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!QCSb!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F407229e9-8a79-4e20-b8a5-1244dc7ba863_1300x848.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!QCSb!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F407229e9-8a79-4e20-b8a5-1244dc7ba863_1300x848.png 424w, https://substackcdn.com/image/fetch/$s_!QCSb!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F407229e9-8a79-4e20-b8a5-1244dc7ba863_1300x848.png 848w, https://substackcdn.com/image/fetch/$s_!QCSb!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F407229e9-8a79-4e20-b8a5-1244dc7ba863_1300x848.png 1272w, https://substackcdn.com/image/fetch/$s_!QCSb!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F407229e9-8a79-4e20-b8a5-1244dc7ba863_1300x848.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!QCSb!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F407229e9-8a79-4e20-b8a5-1244dc7ba863_1300x848.png" width="1300" height="848" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/407229e9-8a79-4e20-b8a5-1244dc7ba863_1300x848.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:848,&quot;width&quot;:1300,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:209687,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.ankursnewsletter.com/i/171599148?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F407229e9-8a79-4e20-b8a5-1244dc7ba863_1300x848.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!QCSb!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F407229e9-8a79-4e20-b8a5-1244dc7ba863_1300x848.png 424w, https://substackcdn.com/image/fetch/$s_!QCSb!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F407229e9-8a79-4e20-b8a5-1244dc7ba863_1300x848.png 848w, https://substackcdn.com/image/fetch/$s_!QCSb!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F407229e9-8a79-4e20-b8a5-1244dc7ba863_1300x848.png 1272w, https://substackcdn.com/image/fetch/$s_!QCSb!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F407229e9-8a79-4e20-b8a5-1244dc7ba863_1300x848.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><strong>Governance debt</strong> is the common thread. Teams rushed to showcase GPT-5 demos, skipped model evals, and now scramble for red-team coverage. Some CIOs ask if they should <em>switch</em> back to smaller <em>models</em> to cut <em>spend</em>; that would be a <em>fixed</em>-cost response to a policy gap, not a technical answer. My suggestion is to embed eval checkpoints per release, store only essential <em>prompt</em> fragments, and budget for guardrail tuning just as you budget for GPUs.</p><p><a href="https://builtin.com/artificial-intelligence/openai-chatgpt-5-hype">GPT-5 is capable, not omnipotent</a>. Its 2024 <em>knowledge cut off</em> still surfaces; I have seen &#8220;free&#8221; advice on SEC rules that missed 2025 amendments. Treat every <em>answer</em> as a first draft, route high-impact calls through deterministic <em>tools</em>, and keep a human reviewer in the loop. With that discipline, the early productivity wins can outweigh the very real operational friction.</p><h2>Cost Realities</h2><p>OpenAI lists <strong>GPT-5 inference at $0.0028 per 1 k output tokens</strong>, almost twice GPT-4o-Mini&#8217;s $0.0015, yet early-access pilots show a 40 percent higher task-completion rate. In a blended model, the <em>right model</em> finishes the <em>job</em> in fewer calls, so total <em>spend</em> drops 11 percent in my Amgen benchmark. The table below captures the main cost drivers:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!LwuX!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3635751a-df43-462a-99d3-65404ad6b47b_1272x642.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!LwuX!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3635751a-df43-462a-99d3-65404ad6b47b_1272x642.png 424w, https://substackcdn.com/image/fetch/$s_!LwuX!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3635751a-df43-462a-99d3-65404ad6b47b_1272x642.png 848w, https://substackcdn.com/image/fetch/$s_!LwuX!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3635751a-df43-462a-99d3-65404ad6b47b_1272x642.png 1272w, https://substackcdn.com/image/fetch/$s_!LwuX!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3635751a-df43-462a-99d3-65404ad6b47b_1272x642.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!LwuX!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3635751a-df43-462a-99d3-65404ad6b47b_1272x642.png" width="1272" height="642" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/3635751a-df43-462a-99d3-65404ad6b47b_1272x642.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:642,&quot;width&quot;:1272,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:145552,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.ankursnewsletter.com/i/171599148?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3635751a-df43-462a-99d3-65404ad6b47b_1272x642.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!LwuX!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3635751a-df43-462a-99d3-65404ad6b47b_1272x642.png 424w, https://substackcdn.com/image/fetch/$s_!LwuX!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3635751a-df43-462a-99d3-65404ad6b47b_1272x642.png 848w, https://substackcdn.com/image/fetch/$s_!LwuX!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3635751a-df43-462a-99d3-65404ad6b47b_1272x642.png 1272w, https://substackcdn.com/image/fetch/$s_!LwuX!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3635751a-df43-462a-99d3-65404ad6b47b_1272x642.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>A total-cost-of-ownership projection for an M365 tenant that touches 20 percent of worker hours shows breakeven in nine months when compared with a GPT-4o-Mini and manual workflow mix. The model assumes 800 k API calls per month, average 4 k-token <em>input</em>, and a 2 percent false-retry rate. If usage is more <em>everyday</em> chat and less structured <em>excel</em> or BI generation, the curve flattens because <em>advanced reasoning</em> is not fully exploited.</p><p><strong>Fine-tuning deserves caution. Selective Adapters look cheap, but I burned 14 percent of the annual inference budget re-training a &#8220;smart contract&#8221; variant that only improved precision by 3 points. Always run an A/B against a prompt-engineering baseline before you </strong><em><strong>switch</strong></em><strong>.</strong></p><p>CFOs who are <em>waiting</em> for &#8220;one fixed price&#8221; will be disappointed. GPT-5&#8217;s economics reward focused, high-value workloads; vague <em>ideas</em> and exploratory <em>stuff</em> should stay on lighter <em>models</em>.</p><h2>Security &amp; Privacy Observations</h2><p>GPT-5 ships with <strong>guardrail embeddings</strong> that flag jailbreak patterns at a 0.4 percent false-positive rate, better than any <em>model</em> I have measured. In red-team drills, only 3 of 500 crafted <em>prompts</em> bypassed policy, versus 27 on GPT-4o. That said, the guardrail occasionally blocks innocuous financial-ratio queries because of string overlap with blocked terms; <em>users</em> need a human override path.</p><p>Apple&#8217;s split-inference design keeps sensitive PII on the device, handing off &gt;4 k-token calls to cloud GPT-5 through a Private Relay tunnel. Regulators point to this as an <em>example</em> architecture for health data. I see a 250 ms latency penalty, but the privacy gain is concrete.</p><p><strong>Security checklist I give </strong><em><strong>customers</strong></em><strong>:</strong></p><ol><li><p>Treat GPT-5 endpoints like any SaaS <em>access</em> point. Enable SCIM provisioning, rotate keys, and keep default scopes limited.<br></p></li><li><p>Layer prompt screening (pre-input) with output DLP (post-response). Dual control cuts both exfiltration and toxic content risk.<br></p></li><li><p>Store only hashed or truncated <em>documents</em> in your log pipeline; full text is rarely needed for incident response.<br></p></li><li><p>Embed deterministic <em>tools</em> for financial approvals so that no free-text answer can move money automatically.</p></li></ol><p>GPT-5 is not inherently unsafe, but its <em>power</em> magnifies small misconfigurations. A Zero-Trust perimeter plus guardrail tuning is cheaper than a breach, even if it adds 5 percent to up-front <em>effort</em>. With that discipline, teams can unlock the <em>productivity</em> gains without handing away the keys to the castle.</p><h2><strong>Architectural Patterns Emerging</strong></h2><p>The year&#8217;s <em>early-access</em> rollouts reveal five repeatable blueprints that make OpenAI GPT 5 workable in production.</p><ul><li><p><strong>Retrieval-Generated Actions (RGA)<br></strong> A query enters, the system retrieves vector-indexed <em>documents</em>, generates a multi-step <em>plan</em>, then calls deterministic <em>tools</em> through a REST <em>API</em>. In Copilot, this loop answers 92% of &#8220;build an <em>Excel</em> pivot&#8221; requests in one pass, trimming human re-work on formatting <em>tasks</em> by 27%. The pattern keeps <em>output tokens</em> low because execution logic happens outside the <em>model</em>.</p></li><li><p><strong>Hierarchical reactive + planning agents<br></strong> One thin &#8220;reactor&#8221; handles short contexts; a &#8220;planner&#8221; with full 256 k <em>input</em> manages <em>complexity</em>. Bench tests show a 15% latency hit versus a single giant agent but a 19% lift in <em>reasoning capabilities</em> on chained instructions.</p></li><li><p><strong>Speculative twin decoding<br></strong> Microsoft&#8217;s Orion clusters run a cheap draft decoder beside GPT-5, then validate. The twin path cuts median response time from 1.1 s to 750 ms for 12 k-token prompts, in effect giving <em>users</em> near-instant <em>answers</em> without extra <em>spend</em>.</p></li><li><p><strong>Hybrid edge / cloud inference<br></strong> Apple&#8217;s Ajax-M4 keeps PII on device and ships heavy lifts to GPT-5. Latency rises 250 ms yet audit flags fall 34%. This split shows regulators a concrete <em>example</em> of privacy by design.</p></li><li><p><strong>Parameter-efficient per-department adapters<br></strong> Selective Adapters (15% weights) let each business unit own a tuned head. Marketing can <em>write</em> web copy while Finance gates on numeric <em>logic</em>. The approach uses 60% less VRAM than full fine-tunes, freeing scarce GPUs for <em>other models</em>.</p></li></ul><h2>Recommendations for CTOs</h2><p>The <em>right model</em> is only half the battle; disciplined engineering closes the gap between demo and KPI.</p><ol><li><p><strong>Limit scope, test hard.</strong> Copy Amgen&#8217;s protocol QC. Pick one narrow domain, build a <em>table</em> of eval metrics, and block launch if <em>advanced reasoning</em> falls below baseline.<br></p></li><li><p><strong>Forecast capacity.</strong> Order GPUs at least six months ahead. For &#8804;8 k-token calls, off-load to CPU inference or a slim SLM to keep <em>cost</em> predictable.<br></p></li><li><p><strong>Automate guardrails.</strong> Bake red-team <em>prompts</em> into CI/CD and demand a 95% Guardrail catch rate before each deploy. Anything less invites governance debt.<br></p></li><li><p><strong>Instrument everything.</strong> Log <em>input</em> and <em>output tokens</em>, but store only hashed text to cut SIEM fees. Use the upcoming OpenAI Governance <em>API</em> to off-load policy checks once it moves from beta.<br></p></li><li><p><strong>Control fine-tuning spend.</strong> Run an A/B against prompt-engineering first; only <em>switch</em> to Selective Adapters if the delta is &#8805;4-point accuracy.<br></p></li><li><p><strong>Educate staff.</strong> Provide <em>developers</em> with fixed &#8220;function-call&#8221; templates. Casual, <em>vague</em> chat at the keyboard is a budget leak.</p></li></ol><p>With these steps, teams can <em>accomplish</em> real gains without writing a blank check.</p><div><hr></div><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!-kh9!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!-kh9!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!-kh9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png" width="1456" height="246" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:246,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:172293,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:&quot;&quot;,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!-kh9!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p>I also host an AI podcast and content series called &#8220;Pioneers.&#8221; This series takes you on an enthralling journey into the minds of AI visionaries, founders, and CEOs who are at the forefront of innovation through AI in their organizations.</p><p>To learn more, please visit <a href="https://www.multimodal.dev/blog">Pioneers</a> on Beehiiv.</p><div><hr></div><h2>What to Watch Next</h2><ul><li><p><strong>Non-English small adapters.</strong> Rumor: OpenAI will ship language-specific heads that cut fine-tune <em>pricing</em> by 60%. If true, GPT-5 could replace boutique local <em>models</em> in LATAM and MENA markets overnight.<br></p></li><li><p><strong>Azure SLM fallback router.</strong> A new policy engine will automatically down-grade <em>everyday</em> prompts to a small language model. Expect 20% lower blended <em>cost</em> for M365 <em>customers</em> who accept the default.<br></p></li><li><p><strong>EU AI Act enforcement Q4 2025.</strong> The first rulings will target autonomous decision agents. Firms that log <em>logic</em> trees and human overrides should pass; those that let the <em>model</em> move money automatically will not.</p><p></p></li><li><p><strong>Open-weight contenders.</strong> Mistral&#8217;s &#8220;Gaillac-10T&#8221; is slated for September. Early leaks put its <em>reasoning</em> within 5% of GPT-5 on AgentBench, though its 128 k context is half as large. If the license is Apache-2, expect rapid <em>developer</em> uptake.<br></p></li><li><p><strong>Knowledge-cut-off extensions.</strong> OpenAI is testing a real-time update stream that patches post-2024 facts through the <em>real time router</em>. If stable, this could reduce current gaps in legal and tax domains by 70%.</p></li></ul><p>Stay alert; each of these shifts can upend your architecture or your budget with little notice.</p><p>GPT-5 sets a <strong>new bar</strong>: enterprises that treat it as merely &#8220;GPT-4 but bigger&#8221; will overspend and under-govern. The winners will be those who design around <em>actions</em>, <em>guardrails</em>, and <em>hybrid infra</em>&#8212;turning a smarter LLM into a safer, cheaper, and truly autonomous teammate.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.ankursnewsletter.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">If you liked reading this post, consider subscribing for more AI content.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Comparing AI Cloud Providers in 2025: Coreweave, Lambda, Cerebras, Etched, Modal, Foundry and New Entrants]]></title><description><![CDATA[Read for a thorough comparison of cloud AI providers like Coreweave, Lambda, Cerebras, Etched, Modal, Foundry, and some new entrants in this space.]]></description><link>https://www.ankursnewsletter.com/p/comparing-ai-cloud-providers-in-2025</link><guid isPermaLink="false">https://www.ankursnewsletter.com/p/comparing-ai-cloud-providers-in-2025</guid><dc:creator><![CDATA[Ankur A. Patel]]></dc:creator><pubDate>Mon, 21 Jul 2025 17:51:58 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!bvv9!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdb1575e9-51ce-4d0b-9e0e-b9dd9db0af61_1200x600.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!bvv9!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdb1575e9-51ce-4d0b-9e0e-b9dd9db0af61_1200x600.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!bvv9!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdb1575e9-51ce-4d0b-9e0e-b9dd9db0af61_1200x600.png 424w, https://substackcdn.com/image/fetch/$s_!bvv9!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdb1575e9-51ce-4d0b-9e0e-b9dd9db0af61_1200x600.png 848w, https://substackcdn.com/image/fetch/$s_!bvv9!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdb1575e9-51ce-4d0b-9e0e-b9dd9db0af61_1200x600.png 1272w, https://substackcdn.com/image/fetch/$s_!bvv9!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdb1575e9-51ce-4d0b-9e0e-b9dd9db0af61_1200x600.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!bvv9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdb1575e9-51ce-4d0b-9e0e-b9dd9db0af61_1200x600.png" width="1200" height="600" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/db1575e9-51ce-4d0b-9e0e-b9dd9db0af61_1200x600.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:600,&quot;width&quot;:1200,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:371623,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.ankursnewsletter.com/i/168877011?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdb1575e9-51ce-4d0b-9e0e-b9dd9db0af61_1200x600.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!bvv9!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdb1575e9-51ce-4d0b-9e0e-b9dd9db0af61_1200x600.png 424w, https://substackcdn.com/image/fetch/$s_!bvv9!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdb1575e9-51ce-4d0b-9e0e-b9dd9db0af61_1200x600.png 848w, https://substackcdn.com/image/fetch/$s_!bvv9!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdb1575e9-51ce-4d0b-9e0e-b9dd9db0af61_1200x600.png 1272w, https://substackcdn.com/image/fetch/$s_!bvv9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdb1575e9-51ce-4d0b-9e0e-b9dd9db0af61_1200x600.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.ankursnewsletter.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Subscribe today to join our community of 2010 AI enthusiasts and business leaders for more insider takes and practical advice on implementing AI in business.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h1><strong>Key Takeaways</strong></h1><ol><li><p>CoreWeave and Lambda Labs lead in <strong>GPU-accelerated AI training and inference</strong>, offering top-tier hardware like NVIDIA Blackwell and B200 with flexible cost-effective scaling.</p></li><li><p>Cerebras delivers unmatched performance for <strong>large-scale LLMs and scientific computing</strong> with its Wafer-Scale Engine (WSE), ideal for highly parallel workloads.</p></li><li><p>Modal and Foundry are <strong>redefining developer experience</strong> with serverless and real-time agent-based platforms that abstract away infrastructure complexity.</p></li><li><p>Etched&#8217;s transformer-specific Sohu ASIC dramatically <strong>reduces energy and hardware needs for LLM inference</strong>, but is only suited for transformer-based models.</p></li><li><p>Choosing the right provider depends on <strong>matching AI workload types (training, inference, deployment) with each platform&#8217;s strengths in performance, pricing, and specialization.</strong></p></li></ol><div><hr></div><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!R9ua!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!R9ua!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!R9ua!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png" width="1456" height="246" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/bea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:246,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:273651,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:&quot;&quot;,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!R9ua!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1456w" sizes="100vw" loading="lazy" fetchpriority="high"></picture><div></div></div></a></figure></div><p>In 2023, I started Multimodal, a Generative AI company that helps organizations automate complex, knowledge-based workflows using AI Agents. Check it out <a href="https://www.multimodal.dev/">here</a>.</p><div><hr></div><p>The AI cloud landscape has rapidly evolved, with both hyperscalers and specialized AI cloud providers pushing the boundaries of scalable, high-performance compute. Below, you'll find a current, detailed comparison of the top dedicated AI cloud platforms&#8212;CoreWeave, Lambda, Cerebras, Etched, Modal, and additional notable entrants&#8212;covering their features, performance, and best-suited use cases for enterprise AI.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!QHEj!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2d219e83-1a22-48a2-b1c8-a55258b19fdc_1284x1242.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!QHEj!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2d219e83-1a22-48a2-b1c8-a55258b19fdc_1284x1242.png 424w, https://substackcdn.com/image/fetch/$s_!QHEj!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2d219e83-1a22-48a2-b1c8-a55258b19fdc_1284x1242.png 848w, https://substackcdn.com/image/fetch/$s_!QHEj!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2d219e83-1a22-48a2-b1c8-a55258b19fdc_1284x1242.png 1272w, https://substackcdn.com/image/fetch/$s_!QHEj!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2d219e83-1a22-48a2-b1c8-a55258b19fdc_1284x1242.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!QHEj!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2d219e83-1a22-48a2-b1c8-a55258b19fdc_1284x1242.png" width="1284" height="1242" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/2d219e83-1a22-48a2-b1c8-a55258b19fdc_1284x1242.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1242,&quot;width&quot;:1284,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:288714,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.ankursnewsletter.com/i/168877011?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2d219e83-1a22-48a2-b1c8-a55258b19fdc_1284x1242.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!QHEj!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2d219e83-1a22-48a2-b1c8-a55258b19fdc_1284x1242.png 424w, https://substackcdn.com/image/fetch/$s_!QHEj!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2d219e83-1a22-48a2-b1c8-a55258b19fdc_1284x1242.png 848w, https://substackcdn.com/image/fetch/$s_!QHEj!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2d219e83-1a22-48a2-b1c8-a55258b19fdc_1284x1242.png 1272w, https://substackcdn.com/image/fetch/$s_!QHEj!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2d219e83-1a22-48a2-b1c8-a55258b19fdc_1284x1242.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h2><strong>1. CoreWeave</strong></h2><ul><li><p><strong>Overview:</strong> CoreWeave leads among AI-first clouds by offering rapid resource scaling (instances in seconds), a rich ecosystem of the latest NVIDIA GPUs (including Blackwell series), and advanced networking (NVIDIA Quantum InfiniBand).</p></li><li><p><strong>Recent Upgrades:</strong> In 2025, became the first to offer NVIDIA RTX PRO 6000 Blackwell Server Edition at scale. MLPerf results show up to 5.6x faster LLM inference, and a $6B PA data center expansion solidifies its U.S. infrastructure.</p></li><li><p><strong>Applications:</strong> Ideal for LLM training, VFX rendering, financial and scientific simulations, and generative AI at massive scale.</p></li><li><p><strong>Cost:</strong> Up to 80% cheaper than general-purpose clouds, with flexible usage-based billing.</p></li></ul><h2><strong>2. Lambda Labs</strong></h2><ul><li><p><strong>Overview:</strong> Lambda specializes in developer-friendly GPU clouds for LLM training, inference, and open-source deployments. Its SOC 2 Type II compliance appeals to enterprise users.</p></li><li><p><strong>Recent Upgrades:</strong> Raised $480M in 2025, expanded to liquid-cooled AI data centers supporting NVIDIA Blackwell/Ultra. Launched 1-Click Clusters for seamless, on-demand scale, and Lambda Inference API for hosted LLMs.</p></li><li><p><strong>Applications:</strong> Best for startups and R&amp;D teams training/fine-tuning large models, generative AI development, and enterprise-scale inference.</p></li><li><p><strong>Cost:</strong> Transparent pricing (A100: $1.25/hr, H100: $2.49/hr) with &#8220;one GPU per user&#8221; ethos.</p></li></ul><h2><strong>3. Cerebras</strong></h2><ul><li><p><strong>Overview:</strong> Famous for the Wafer-Scale Engine (WSE), Cerebras packs unparalleled compute density for AI&#8212;2.6T+ transistors and 850,000 AI cores in a single chip.</p></li><li><p><strong>Recent Upgrades:</strong> Deployed 6 new data centers (North America/France), launched Qwen3-235B (frontier LLM) at 1/10th the cost of closed models, deep integration with Hugging Face for inference hosting.</p></li><li><p><strong>Applications:</strong> Unmatched for LLM and scientific simulation, medical/drug discovery, and workloads requiring fast, massive inference and modeling.</p></li><li><p><strong>Differentiator:</strong> Efficient mixture-of-experts LLMs, high context window (up to 131K), deep cost savings for production deployment.</p></li></ul><h2><strong>4. Modal</strong></h2><ul><li><p><strong>Overview:</strong> Modal eliminates DevOps headaches with serverless, containerized, GPU-backed compute&#8212;ideal for organizations that want to focus on code, not infrastructure.</p></li><li><p><strong>Recent Upgrades:</strong> &gt;100 enterprise clients, dynamic scaling driven by agents, Rust backend for rapid, pre-configured launches.</p></li><li><p><strong>Applications:</strong> Best for rapid generative AI inference, computational biotech workloads, automated transcription, and batch analytics jobs.</p></li><li><p><strong>Cost:</strong> Pay-as-you-go usage model with aggressive optimization for job cost and latency.</p></li></ul><h2><strong>5. Etched</strong></h2><ul><li><p><strong>Overview:</strong> Etched's Sohu chip is a transformer-specific ASIC, achieving ultra-high performance for LLM inference&#8212;up to 500,000 tokens/sec for Llama-70B&#8212;which can replace up to 160 H100s in one 8xSohu box.</p></li><li><p><strong>Recent Upgrades:</strong> Entered the market in 2025, addressing high energy use by leveraging specialization; &gt;90% FLOPS utilization.</p></li><li><p><strong>Applications:</strong> Only for transformer-based inference&#8212;chatbots, real-time NLP, and high-throughput services.</p></li><li><p><strong>Cost &amp; Efficiency:</strong> Large energy and space savings; only worth it for workloads that are 100% transformer-aligned.</p></li></ul><h2><strong>6. Foundry (Microsoft Azure AI Foundry)</strong></h2><ul><li><p><strong>Overview:</strong> Microsoft&#8217;s Azure AI Foundry is a &#8220;real-time compute market&#8221; and LLM ops platform, tightly integrated with Azure cloud.</p></li><li><p><strong>Recent Upgrades:</strong> Supports Grok and advanced models, AI Agent workflow tools, robust monitoring, and the Model Router that auto-selects the best model for each query. Dramatically improved developer tooling in 2025.</p></li><li><p><strong>Applications:</strong> Real-time LLM training/inference, agent-based enterprise automation, hyperparameter tuning, industry cloud integrations.</p></li><li><p><strong>Differentiator:</strong> Multi-agent orchestration, robust agent lifecycle management, deep integration with Microsoft&#8217;s enterprise tech stack.</p></li></ul><h2><strong>Additional Notable Players </strong></h2><ul><li><p><strong>Spectro Cloud:</strong> Rapidly rising in composable infrastructure for AI workloads, appearing alongside Lambda and CoreWeave as a high-growth upstart.</p></li><li><p><strong>Cloudera, AWS (Trainium/Inferentia), Google Cloud (Gemini):</strong> Remain strong but less differentiated on AI-specific hardware&#8212;however, provide best-in-class multicloud integration and enterprise support.</p></li><li><p><strong>Groq, Perceive:</strong> Newcomers focusing on transformer-optimized and low-energy AI inference chips emerging as potential Etched competitors.</p></li></ul><h2><strong>Market and Technology Trends in 2025</strong></h2><ul><li><p><strong>AI hardware diversity:</strong> Blackwell, WSE, B200, Sohu, and cloud ASICs are driving faster, cheaper, and greener AI infrastructure.</p></li><li><p><strong>Specialization pays:</strong> Transformer-specific platforms (Etched, Groq) are converging on the inference boom; generic clouds differentiate with scale and integration.</p></li><li><p><strong>Enterprise shift:</strong> SOC2/ISO compliance and direct LLM inference APIs are now must-haves for enterprise adoption.</p></li><li><p><strong>Cloud spend soars:</strong> AI now drives an estimated half of new cloud revenue, with spending expected to surpass $723B globally in 2025.</p></li><li><p><strong>Agent-driven ops:</strong> Serverless and agent-based operation models (Modal, Foundry) slash dev time and reduce infrastructure friction.</p></li></ul><div><hr></div><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!-kh9!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!-kh9!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!-kh9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png" width="1456" height="246" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:246,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:172293,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:&quot;&quot;,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!-kh9!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p>I also host an AI podcast and content series called &#8220;Pioneers.&#8221; This series takes you on an enthralling journey into the minds of AI visionaries, founders, and CEOs who are at the forefront of innovation through AI in their organizations.</p><p>To learn more, please visit <a href="https://www.multimodal.dev/blog">Pioneers</a> on Beehiiv.</p><div><hr></div><h2><strong>How to Choose the Right Provider</strong></h2><ul><li><p><strong>Match your workload:</strong> LLM training or VFX = CoreWeave/Lambda; transformer inference = Etched; fast, serverless deployment = Modal; massive parallel compute = Cerebras; integrated enterprise workflows = Foundry.</p></li><li><p><strong>Consider cost vs. performance:</strong> CoreWeave and Lambda offer leading price/performance for GPU-heavy jobs.</p></li><li><p><strong>Future-proofing:</strong> Look for rapid hardware releases (Blackwell, new ASICs), ecosystem fit, and compliance capabilities.</p></li><li><p><strong>Integration:</strong> Consider how well a provider fits into your existing stack, especially if leveraging Azure/Microsoft or hybrid cloud models.</p></li></ul><p>AI-first clouds are complementing (and sometimes outpacing) the hyperscalers with rapid hardware innovation and a laser focus on high-performance AI workloads&#8212;a trend that is only expected to accelerate throughout 2025 and beyond.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.ankursnewsletter.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">If you liked reading this post, consider subscribing for more AI content.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[What Are AI Evals And Why Are They a Technical Moat]]></title><description><![CDATA[AI evals are the enterprise moat: systematic evaluation systems that transform probabilistic AI into reliable assets. Learn more about what they are and how to work with them.]]></description><link>https://www.ankursnewsletter.com/p/what-are-ai-evals-and-why-are-they</link><guid isPermaLink="false">https://www.ankursnewsletter.com/p/what-are-ai-evals-and-why-are-they</guid><dc:creator><![CDATA[Ankur A. Patel]]></dc:creator><pubDate>Thu, 03 Jul 2025 19:59:48 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!vRwg!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6ec32a8-23ed-4da9-bf89-c76843d79295_1200x600.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!vRwg!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6ec32a8-23ed-4da9-bf89-c76843d79295_1200x600.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!vRwg!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6ec32a8-23ed-4da9-bf89-c76843d79295_1200x600.png 424w, https://substackcdn.com/image/fetch/$s_!vRwg!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6ec32a8-23ed-4da9-bf89-c76843d79295_1200x600.png 848w, https://substackcdn.com/image/fetch/$s_!vRwg!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6ec32a8-23ed-4da9-bf89-c76843d79295_1200x600.png 1272w, https://substackcdn.com/image/fetch/$s_!vRwg!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6ec32a8-23ed-4da9-bf89-c76843d79295_1200x600.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!vRwg!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6ec32a8-23ed-4da9-bf89-c76843d79295_1200x600.png" width="1200" height="600" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e6ec32a8-23ed-4da9-bf89-c76843d79295_1200x600.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:600,&quot;width&quot;:1200,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:375108,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.ankursnewsletter.com/i/167467555?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6ec32a8-23ed-4da9-bf89-c76843d79295_1200x600.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!vRwg!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6ec32a8-23ed-4da9-bf89-c76843d79295_1200x600.png 424w, https://substackcdn.com/image/fetch/$s_!vRwg!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6ec32a8-23ed-4da9-bf89-c76843d79295_1200x600.png 848w, https://substackcdn.com/image/fetch/$s_!vRwg!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6ec32a8-23ed-4da9-bf89-c76843d79295_1200x600.png 1272w, https://substackcdn.com/image/fetch/$s_!vRwg!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6ec32a8-23ed-4da9-bf89-c76843d79295_1200x600.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.ankursnewsletter.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Subscribe today to join our community of 2010 AI enthusiasts and business leaders for more insider takes and practical advice on implementing AI in business.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h2>Key Takeaways</h2><ol><li><p><strong>Evals surpass models as moats</strong> by converting probabilistic outputs into auditable, reliable assets for enterprise deployment.</p></li><li><p><strong>Tiered evaluation rigor</strong> (L1-L4) enables progressive validation from syntax checks to business impact simulation.</p></li><li><p><strong>Hybrid scoring systems</strong> combine automated metrics, human review, and LLM judges for compliance-critical use cases.</p></li><li><p><strong>Probabilistic methods</strong> (Monte Carlo, stability scoring) address non-determinism in high-stakes environments.</p></li><li><p><strong>Evaluation-aware MLOps</strong> integrates evals into CI/CD pipelines to enforce governance and enable continuous improvement.</p></li></ol><div><hr></div><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!R9ua!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!R9ua!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!R9ua!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png" width="1456" height="246" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/bea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:246,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:273651,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:&quot;&quot;,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!R9ua!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1456w" sizes="100vw" loading="lazy" fetchpriority="high"></picture><div></div></div></a></figure></div><p>In 2023, I started Multimodal, a Generative AI company that helps organizations automate complex, knowledge-based workflows using AI Agents. Check it out <a href="https://www.multimodal.dev/">here</a>.</p><div><hr></div><p>The non-deterministic nature of AI fundamentally disrupts traditional software paradigms. Unlike deterministic systems where unit tests verify fixed outputs, generative AI produces probabilistic results. A single input prompt can yield divergent outputs across model versions, rendering conventional QA inadequate. Binary pass/fail checks fail to capture nuance in natural language outputs or complex completion functions. Industry leaders emphasize this shift:</p><blockquote><p><strong><a href="https://x.com/garrytan/status/1892952656940880036">Garry Tan (YC CEO)</a></strong>: "AI evals are emerging as the real moat for AI startups."</p></blockquote><p>Evals operationalize reliability through systematic evaluation. They transform subjective assessments into quantifiable metrics, tracking performance across failure modes, model versions, and production data. For enterprise AI applications, this isn&#8217;t just testing. It&#8217;s the core process that gates deployment, informs fine-tuning, and turns probabilistic systems into trusted assets.</p><p>Let&#8217;s learn more about evals and how they can help build robust enterprise AI systems.</p><h2>Deconstructing Evals: Beyond Basic Testing</h2><h3>What Evals Actually Measure</h3><p>Enterprise AI evals transcend traditional unit tests by quantifying four critical dimensions:</p><ul><li><p><strong>Accuracy vs. precision tradeoffs</strong>: Measuring factual correctness while allowing for contextual nuance in natural language outputs.</p></li><li><p><strong>Contextual alignment</strong>: Verifying outputs comply with business rules, regulatory constraints, and domain-specific logic.</p></li><li><p><strong>Safety &amp; bias quantification</strong>: Detecting toxicity, discrimination, and security vulnerabilities in model behavior.</p></li><li><p><strong>Cost-performance optimization</strong>: Balancing inference cost against quality thresholds for sustainable scaling.</p></li></ul><p>Unlike binary unit tests, systematic evaluation analyzes how different model versions handle edge cases in production data. This requires creating high quality evals that simulate real user input and failure modes.</p><h3>The Evaluation Taxonomy</h3><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!-mRv!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5dd6b963-f60d-469d-87df-6d77a877ff0b_746x220.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!-mRv!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5dd6b963-f60d-469d-87df-6d77a877ff0b_746x220.png 424w, https://substackcdn.com/image/fetch/$s_!-mRv!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5dd6b963-f60d-469d-87df-6d77a877ff0b_746x220.png 848w, https://substackcdn.com/image/fetch/$s_!-mRv!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5dd6b963-f60d-469d-87df-6d77a877ff0b_746x220.png 1272w, https://substackcdn.com/image/fetch/$s_!-mRv!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5dd6b963-f60d-469d-87df-6d77a877ff0b_746x220.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!-mRv!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5dd6b963-f60d-469d-87df-6d77a877ff0b_746x220.png" width="746" height="220" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/5dd6b963-f60d-469d-87df-6d77a877ff0b_746x220.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:220,&quot;width&quot;:746,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!-mRv!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5dd6b963-f60d-469d-87df-6d77a877ff0b_746x220.png 424w, https://substackcdn.com/image/fetch/$s_!-mRv!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5dd6b963-f60d-469d-87df-6d77a877ff0b_746x220.png 848w, https://substackcdn.com/image/fetch/$s_!-mRv!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5dd6b963-f60d-469d-87df-6d77a877ff0b_746x220.png 1272w, https://substackcdn.com/image/fetch/$s_!-mRv!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5dd6b963-f60d-469d-87df-6d77a877ff0b_746x220.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p>Effective evaluation systems combine techniques based on risk tolerance and use case. When running evals, AI PMs should:</p><ol><li><p><strong>Define test cases</strong> covering critical failure modes</p></li><li><p><strong>Generate synthetic data</strong> for edge scenarios</p></li><li><p><strong>Write prompts</strong> using eval templates (e.g., OpenAI Evals format)</p></li><li><p><strong>Analyze metrics</strong> like hallucination rates and compliance scores</p></li></ol><p>For example, using' pip install evals' establishes a baseline, while fine-tuning completion functions against domain-specific JSON datasets elevates precision. The evaluation process culminates in a final report comparing model versions against business KPIs.</p><h2>The Technical Anatomy of Enterprise-Grade Evals</h2><h3>Component Engineering</h3><h4>Prompt Conditioning</h4><p>Enterprise eval templates standardize input prompts with contextual guardrails. This ensures consistent test cases across model versions while isolating variables during evaluation.</p><h4>Constraint Embedding</h4><p>Critical for compliance-heavy AI applications:</p><ul><li><p>Regulatory clauses &#8594; Vectorized embeddings for semantic matching</p></li><li><p>Business rules &#8594; Finite state machines enforcing decision trees<br> Transforms abstract policies into executable code for systematic evaluation.</p></li></ul><h4>Failure Mode Instrumentation</h4><p> Proactive detection systems:</p><ul><li><p>Hallucination detectors: Trigger alerts when entropy exceeds domain-specific thresholds</p></li><li><p>Drift sensors: Monitor KL-divergence between training and production data<br> Enables real-time intervention before issues impact business processes.</p></li></ul><h3>Evaluation Rigor Levels</h3><p>A tiered framework for creating high quality evals:</p><ul><li><p><strong>L1</strong>: Basic output structure (JSON format, key presence)</p></li><li><p><strong>L2</strong>: Meaning accuracy (factual correctness, no contradictions)</p></li><li><p><strong>L3</strong>: Domain alignment (compliance, brand voice, safety)</p></li><li><p><strong>L4</strong>: Outcome simulation (cost/benefit analysis of AI-generated decisions)</p></li></ul><p>When running evals, this progression allows machine learning engineers to:</p><ol><li><p>Start with automated evaluation of syntax (L1)</p></li><li><p>Progress to human evaluation for nuanced tasks (L3)</p></li><li><p>Validate business impact using synthetic data mirroring production environments (L4)<br> The final report should benchmark performance across all levels to guide fine tuning.</p></li></ol><h2>Implementation Framework for Technical Teams</h2><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!KSPn!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70d664f3-e5f8-42f8-98ff-431e653a1082_1270x820.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!KSPn!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70d664f3-e5f8-42f8-98ff-431e653a1082_1270x820.png 424w, https://substackcdn.com/image/fetch/$s_!KSPn!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70d664f3-e5f8-42f8-98ff-431e653a1082_1270x820.png 848w, https://substackcdn.com/image/fetch/$s_!KSPn!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70d664f3-e5f8-42f8-98ff-431e653a1082_1270x820.png 1272w, https://substackcdn.com/image/fetch/$s_!KSPn!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70d664f3-e5f8-42f8-98ff-431e653a1082_1270x820.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!KSPn!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70d664f3-e5f8-42f8-98ff-431e653a1082_1270x820.png" width="1270" height="820" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/70d664f3-e5f8-42f8-98ff-431e653a1082_1270x820.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:820,&quot;width&quot;:1270,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;Mastering AI Evals: A Complete Guide for PMs&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Mastering AI Evals: A Complete Guide for PMs" title="Mastering AI Evals: A Complete Guide for PMs" srcset="https://substackcdn.com/image/fetch/$s_!KSPn!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70d664f3-e5f8-42f8-98ff-431e653a1082_1270x820.png 424w, https://substackcdn.com/image/fetch/$s_!KSPn!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70d664f3-e5f8-42f8-98ff-431e653a1082_1270x820.png 848w, https://substackcdn.com/image/fetch/$s_!KSPn!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70d664f3-e5f8-42f8-98ff-431e653a1082_1270x820.png 1272w, https://substackcdn.com/image/fetch/$s_!KSPn!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70d664f3-e5f8-42f8-98ff-431e653a1082_1270x820.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><a href="https://substackcdn.com/image/fetch/$s_!KSPn!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F70d664f3-e5f8-42f8-98ff-431e653a1082_1270x820.png">Source</a></figcaption></figure></div><h3>The Eval Development Lifecycle</h3><h4>Requirement Decomposition</h4><p>Transform business objectives into executable test cases:</p><ul><li><p>Regulatory compliance &#8594; Convert SEC clauses into verification steps (e.g., "Output must cite &#167;12(b)-1 when discussing fees")</p></li><li><p>SLA targets &#8594; Quantify as eval metrics (e.g., 95% accuracy on financial document parsing)<br> Enables systematic evaluation aligned with business outcomes.</p></li></ul><h4>Test Harness Architecture</h4><p> Robust infrastructure for repeatable evals:</p><ul><li><p><strong>Data versioning</strong>: Track eval dataset iterations with DVC</p></li><li><p><strong>Pipeline orchestration</strong>: Schedule runs via Airflow/Kubeflow</p></li><li><p><strong>Continuous evaluation</strong></p><ul><li><p>Canary testing: Deploy new model versions to 5% traffic</p></li><li><p>Automated regression detection: Flag performance drops using historical benchmarks<br> Ensures eval process integrity across development cycles.</p></li></ul></li></ul><h3>Optimization Techniques</h3><h4>Cost-Efficient Scaling</h4><ul><li><p><strong>Stratified sampling</strong>: Prioritize high-impact test cases (e.g., 80% production data + 20% synthetic edge cases</p></li><li><p><strong>Distilled graders</strong>: Replace GPT-4 evaluators with fine-tuned TinyLlama models after calibration</p></li></ul><h4>Latency Optimization</h4><ul><li><p><strong>Async pipelines</strong>: Decouple execution from scoring (e.g., run evals during off-peak hours)</p></li></ul><h3>Best Practices</h3><ol><li><p><strong>Version control</strong>: Track prompt templates and model versions in Git</p></li><li><p><strong>Automated evaluation</strong>: Use pip install evals for baseline metrics</p></li><li><p><strong>Hybrid validation</strong>: Combine LLM-as-Judge scoring with SME spot checks</p></li><li><p><strong>Failure mode analysis</strong>: Log unexpected outputs to refine test cases</p></li></ol><p>When running evals, this framework enables:</p><ul><li><p>Early detection of regression in new model versions</p></li><li><p>Quantifiable progress tracking via eval metrics (precision/recall/F1)</p></li><li><p>Efficient resource allocation using stratified sampling</p></li><li><p>Auditable final reports for compliance requirements</p></li></ul><p>For mission-critical AI applications, bake these components into CI/CD pipelines. This transforms evaluation from a checkpoint into a continuous improvement engine.</p><h2>Advanced Technical Considerations</h2><h3>Handling Non-Determinism</h3><p>Enterprise AI evals demand probabilistic approaches to address inherent model variability:</p><ul><li><p><strong>Monte Carlo confidence intervals</strong>: Run 100+ iterations of the same prompt to establish output distribution bounds</p></li><li><p><strong>Probabilistic scoring</strong>: Replace binary decisions with likelihood-based metrics (e.g., "85% confidence this output complies with &#167;32(a)")</p></li></ul><h3>Enterprise-Specific Challenges</h3><h4>Data Sovereignty</h4><ul><li><p>Design on-prem eval clusters for air-gapped environments</p></li><li><p>Generate synthetic data using domain-constrained LLMs (e.g., "Create plausible patient records without real PHI")</p></li></ul><h4>Legacy Integration</h4><ul><li><p>Build API shims for mainframe systems using OpenAPI specifications</p></li><li><p>Implement evaluation-aware caching: Store frequent regulatory queries to reduce latency</p></li></ul><h3>Security &amp; Compliance</h3><ul><li><p><strong>PII scrubbing</strong>: Automatically redact sensitive data in eval datasets using transformer-based NER</p></li><li><p><strong>Audit trails</strong>: Log all evaluation process steps in immutable JSON format for ISO 27001 compliance</p></li><li><p><strong>Inversion attack prevention</strong>:</p><ul><li><p>Mask API keys in generated code</p></li><li><p>Restrict output granularity via completion functions</p></li></ul></li></ul><p>When running evals in regulated environments:</p><ol><li><p>Use synthetic data for initial testing phases</p></li><li><p>Validate against production data only after L3 contextual alignment checks</p></li><li><p>Embed compliance rules directly into eval templates (e.g., HIPAA constraints in prompt conditioning)</p></li></ol><p>By combining probabilistic methods with sovereign infrastructure, technical teams can deploy AI applications with quantifiable risk profiles.</p><h2>The Technical Evolution of Evals</h2><h3>Emerging Capabilities</h3><h4>AutoEval Systems</h4><p> Next-generation evaluation systems enable autonomous quality management:</p><ul><li><p><strong>Self-improving test cases</strong>: AI generates iterative evals using failure patterns from production data</p></li><li><p><strong>Causal diagnosis engines</strong>: Pinpoint root causes (e.g., prompt flaws vs. model drift) using Bayesian networks<br>Platforms like AutoEval fine-tune lightweight evaluators (e.g., alBERTa) that learn from LLM-judged interactions, enabling continuous evaluation without manual oversight.</p></li></ul><h4>Multi-Agent Evaluation</h4><p> Adversarial frameworks stress-test AI systems:</p><ul><li><p><strong>Red teaming agents</strong>: Simulate malicious inputs to probe security vulnerabilities</p></li><li><p><strong>Consistency validators</strong>: Cross-check outputs across model versions using ensemble voting<br> This mirrors enterprise red teaming practices where specialized agents jailbreak systems to expose weaknesses before deployment.</p></li></ul><p>These advancements transform evals from static checks to dynamic integrity guardians, where evaluation systems autonomously refine test cases and enforce compliance 24/7.</p><div><hr></div><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!-kh9!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!-kh9!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!-kh9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png" width="1456" height="246" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:246,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:172293,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:&quot;&quot;,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!-kh9!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p>I also host an AI podcast and content series called &#8220;Pioneers.&#8221; This series takes you on an enthralling journey into the minds of AI visionaries, founders, and CEOs who are at the forefront of innovation through AI in their organizations.</p><p>To learn more, please visit <a href="https://www.multimodal.dev/blog">Pioneers</a> on Beehiiv.</p><div><hr></div><h2>Building the Eval-Centric Organization</h2><p>In enterprise AI, <strong>evals surpass model architecture</strong> as the true competitive moat. While models rapidly commoditize, robust evaluation systems deliver lasting advantage by transforming probabilistic outputs into trusted business assets. For technical leaders, three actions are critical:</p><h3>Technical Action Plan</h3><ol><li><p><strong>Establish eval SWAT teams</strong>: Cross-functional units (ML engineers, domain experts, compliance officers) owning the evaluation process end-to-end</p></li><li><p><strong>Implement evaluation-aware MLOps</strong>: Integrate evals into CI/CD pipelines using tools like Kubeflow Evals and Weights &amp; Biases</p></li><li><p><strong>Mandate eval rigor</strong>: Enforce quantitative metrics as deployment gates in AI governance frameworks</p></li></ol><p>The future belongs to organizations treating evaluations as core intellectual property. Advanced eval templates, failure mode databases, and scoring methodologies will become strategic assets. These enable continuous improvement while mitigating emerging risks in production environments. As AI applications evolve, organizations institutionalizing eval-centric development will lead to measurable reliability and controlled innovation.</p><p>I&#8217;ll come back soon with more on building agentic AI for enterprises.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.ankursnewsletter.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">If you liked reading this post, consider subscribing for more AI content.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[How to Build Agentic AI for Enterprises]]></title><description><![CDATA[Building agentic AI systems for enterprises can be an engineering nightmare. Learn what techniques, principles, and methods help.]]></description><link>https://www.ankursnewsletter.com/p/how-to-build-agentic-ai-for-enterprises</link><guid isPermaLink="false">https://www.ankursnewsletter.com/p/how-to-build-agentic-ai-for-enterprises</guid><dc:creator><![CDATA[Ankur A. Patel]]></dc:creator><pubDate>Fri, 13 Jun 2025 20:24:30 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!D5fi!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2728d510-0943-4a60-ada2-c175ab93d1f9_1200x600.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!D5fi!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2728d510-0943-4a60-ada2-c175ab93d1f9_1200x600.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!D5fi!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2728d510-0943-4a60-ada2-c175ab93d1f9_1200x600.png 424w, https://substackcdn.com/image/fetch/$s_!D5fi!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2728d510-0943-4a60-ada2-c175ab93d1f9_1200x600.png 848w, https://substackcdn.com/image/fetch/$s_!D5fi!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2728d510-0943-4a60-ada2-c175ab93d1f9_1200x600.png 1272w, https://substackcdn.com/image/fetch/$s_!D5fi!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2728d510-0943-4a60-ada2-c175ab93d1f9_1200x600.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!D5fi!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2728d510-0943-4a60-ada2-c175ab93d1f9_1200x600.png" width="1200" height="600" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/2728d510-0943-4a60-ada2-c175ab93d1f9_1200x600.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:600,&quot;width&quot;:1200,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:370162,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.ankursnewsletter.com/i/165895342?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2728d510-0943-4a60-ada2-c175ab93d1f9_1200x600.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!D5fi!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2728d510-0943-4a60-ada2-c175ab93d1f9_1200x600.png 424w, https://substackcdn.com/image/fetch/$s_!D5fi!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2728d510-0943-4a60-ada2-c175ab93d1f9_1200x600.png 848w, https://substackcdn.com/image/fetch/$s_!D5fi!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2728d510-0943-4a60-ada2-c175ab93d1f9_1200x600.png 1272w, https://substackcdn.com/image/fetch/$s_!D5fi!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2728d510-0943-4a60-ada2-c175ab93d1f9_1200x600.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.ankursnewsletter.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Subscribe today to join our community of 2000 AI enthusiasts and business leaders for more insider takes and practical advice on implementing AI in business.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h2><strong>Key Takeaways</strong></h2><ol><li><p>Agentic AI replaces stateless ML with stateful architectures using vector DBs/RAG for enterprise-scale context tracking.</p></li><li><p>Multi-agent systems reduce IT incident resolution time by 80% through specialized security/remediation/compliance agents.</p></li><li><p>Supply chain agents balance cost/sustainability via ERP API integration and reinforcement learning conflict resolution.</p></li><li><p>Production systems require governance layers with ABAC controls and circuit breakers for HIPAA/SOC2 compliance.</p></li><li><p>Successful implementation demands phased rollouts: process mining &#8594; bounded pilots &#8594; Zero Trust scaling.</p></li></ol><div><hr></div><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!R9ua!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!R9ua!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!R9ua!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png" width="1456" height="246" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/bea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:246,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:273651,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:&quot;&quot;,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!R9ua!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1456w" sizes="100vw" loading="lazy" fetchpriority="high"></picture><div></div></div></a></figure></div><p>In 2023, I started Multimodal, a Generative AI company that helps organizations automate complex, knowledge-based workflows using AI Agents. Check it out <a href="https://www.multimodal.dev/">here</a>.</p><div><hr></div><p>Building agentic workflows for enterprises comes with several engineering challenges. Everything from model to workflow selection can seem daunting, especially with continuous updates to frameworks and models. In this article, I will break down the key things you should keep in mind if you&#8217;re spearheading an agentic AI initiative in an enterprise.</p><h2>The Agentic Paradigm Shift: Technical Differentiation</h2><p>Agentic AI systems mark a fundamental departure from traditional machine learning pipelines through three core architectural shifts. First, <strong>statefulness</strong> replaces stateless inference - where conventional ML models process isolated data batches, agentic architectures maintain persistent context across interactions. This enables continuity in complex workflows like supply chain management, where autonomous agents track progress across multiple systems while protecting sensitive data.</p><p>Second, <strong>multi-agent orchestration</strong> supersedes monolithic models. OpenAI's framework exemplifies this by separating roles into specialized AI agents (analysis, decision, execution) that collaborate dynamically. Unlike single-model approaches struggling with vast amounts of data, these distributed systems allocate specific tasks to optimized AI models - a security agent handles access controls while a logistics agent processes real-time sensor data.</p><p>Third, the <strong>autonomy spectrum</strong> evolves from rigid rule-based delegation to LLM-driven goal decomposition. While early robotic process automation required constant human input for task definitions, modern agentic systems use reinforcement learning to break high-level objectives into executable steps. This shift enables AI-powered agents to handle complex scenarios like dynamic problem-solving in software development pipelines while maintaining human oversight through circuit breaker protocols.</p><h3>Key Technical Components</h3><p>Modern agentic architectures combine three critical elements:</p><ol><li><p><strong>Hybrid Memory Systems</strong></p><ul><li><p>Context window limitations in large language models are overcome through vector databases + RAG patterns</p></li><li><p>Enables intelligent systems to process enterprise-scale data while tracking progress across interactions</p></li></ul></li><li><p><strong>Expanded Action Spaces</strong></p><ul><li><p>API toolchains integrated with code execution engines</p></li><li><p>Autonomous agents execute tasks ranging from basic interactions to complex processes like automating decision-making in patient data analysis</p></li></ul></li><li><p><strong>Resilience Frameworks</strong></p><ul><li><p>Circuit breakers halt unintended consequences in real-time</p></li><li><p>Human-in-the-loop escalation paths for high-stakes scenarios (financial approvals, medical diagnoses)</p></li></ul></li></ol><p>These components enable agentic AI to streamline software development, optimize supply chains, and improve customer satisfaction while maintaining rigorous testing protocols. By integrating with existing enterprise systems through secure access controls, these architectures balance operational efficiency with robust data protection - a critical advancement over earlier generative AI solutions.</p><h2>High-Value Enterprise Use Cases (Technical Implementation View)</h2><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!-G2y!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4c8e00af-a397-40a4-b5fa-4db1656b7ba8_1024x1024.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!-G2y!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4c8e00af-a397-40a4-b5fa-4db1656b7ba8_1024x1024.png 424w, https://substackcdn.com/image/fetch/$s_!-G2y!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4c8e00af-a397-40a4-b5fa-4db1656b7ba8_1024x1024.png 848w, https://substackcdn.com/image/fetch/$s_!-G2y!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4c8e00af-a397-40a4-b5fa-4db1656b7ba8_1024x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!-G2y!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4c8e00af-a397-40a4-b5fa-4db1656b7ba8_1024x1024.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!-G2y!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4c8e00af-a397-40a4-b5fa-4db1656b7ba8_1024x1024.png" width="1024" height="1024" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/4c8e00af-a397-40a4-b5fa-4db1656b7ba8_1024x1024.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1024,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1316287,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.ankursnewsletter.com/i/165895342?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4c8e00af-a397-40a4-b5fa-4db1656b7ba8_1024x1024.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!-G2y!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4c8e00af-a397-40a4-b5fa-4db1656b7ba8_1024x1024.png 424w, https://substackcdn.com/image/fetch/$s_!-G2y!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4c8e00af-a397-40a4-b5fa-4db1656b7ba8_1024x1024.png 848w, https://substackcdn.com/image/fetch/$s_!-G2y!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4c8e00af-a397-40a4-b5fa-4db1656b7ba8_1024x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!-G2y!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4c8e00af-a397-40a4-b5fa-4db1656b7ba8_1024x1024.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p><h3>1. Autonomous IT Operations</h3><p><strong>Pattern</strong>: Multi-agent incident response teams<br> Modern agentic AI systems deploy specialized AI agents that collaborate to resolve IT incidents without constant human input. For example:</p><ul><li><p><strong>Security Agent</strong>: Scans network traffic using machine learning algorithms to detect anomalies in real-time data streams</p></li><li><p><strong>Remediation Agent</strong>: Executes patch deployments through robotic process automation, enforcing access controls to protect sensitive data</p></li><li><p><strong>Compliance Agent</strong>: Generates audit trails automatically, integrating with existing enterprise systems like SAP/Oracle ERP platforms</p></li></ul><h4>Technical Considerations</h4><ul><li><p><strong>Privileged Access Management</strong>: AI agents operate with least-privilege permissions, restricted to specific tasks like read-only access to patient data or API-based patch deployment</p></li><li><p><strong>Real-Time Observability</strong>: Platforms like <a href="https://sixthsense.rakuten.com/data-observability/blog/Top-5-Data-Observability-Tools-A-Comprehensive-Comparison-for-2025">Rakuten SixthSense</a> provide dynamic monitoring of agent-driven actions across multiple systems, enabling rapid debugging of unintended consequences</p></li><li><p><strong>Case Study</strong>: <a href="https://www.jamf.com/blog/time-savings-you-can-expect-with-jamf/">Jamf Pro&#8217;s autonomous provisioning agent </a>reduces device setup time by <strong>80%</strong> through zero-touch deployment workflows, processing vast amounts of device data while maintaining GDPR compliance</p></li></ul><h3>2. Supply Chain Optimization</h3><p><strong>Pattern</strong>: Continuous multi-objective optimization<br>Agentic AI marks a paradigm shift in supply chain management by deploying intelligent systems that balance competing priorities:</p><ul><li><p><strong>Demand Forecasting Agent</strong>: Analyzes market trends and real-time sensor data from IoT networks</p></li><li><p><strong>Logistics Routing Agent</strong>: Optimizes transportation using utility functions to resolve conflicts between cost, speed, and sustainability</p></li><li><p><strong>Compliance Agent</strong>: Ensures regulatory adherence while processing sensitive supplier data across external systems</p></li></ul><h4><strong>Technical Deep Dive</strong></h4><ul><li><p><strong>ERP Integration</strong>: AI-powered agents interface with SAP/Oracle APIs to synchronize inventory data, automating decision-making for complex workflows</p></li><li><p><strong>Conflict Resolution</strong>: Multi-agent systems apply reinforcement learning to balance objectives (e.g., reducing carbon footprint vs. minimizing costs)</p></li><li><p><strong>Real-World Impact</strong>: <a href="https://c3.ai/blog/transforming-supply-chain-optimization-with-c3-ais-multi-hop-orchestration-agents-part-4/">C3 AI&#8217;s multi-hop orchestration</a> agents enable <strong>15% supply chain cost reductions</strong> through adaptive inventory buffering and risk-weighted routing</p></li></ul><h3>3. AI-Driven Software Development</h3><p><strong>Pattern</strong>: Recursive code improvement loops<br>Autonomous agents are streamlining software development through AI-powered toolchains:</p><ul><li><p><strong>Spec Interpreter</strong>: Converts natural language requirements into technical user stories using large language models</p></li><li><p><strong>Implementation Agent</strong>: Generates production-ready code while maintaining compatibility with existing enterprise systems</p></li><li><p><strong>Security Auditor</strong>: Scans for vulnerabilities in AI-generated code, enforcing access controls for sensitive data processing</p></li><li><p><strong>Test Generator</strong>: Creates rigorous testing protocols using real-time data from CI/CD pipelines</p></li></ul><h4>Implementation Challenges</h4><ul><li><p><strong>Technical Debt Management</strong>: Tools like SonarQube and CodeGuru track progress on code quality, automatically flagging complex processes for refactoring</p></li><li><p><strong>Version Control</strong>: LangChain workflows implement Git-based tracking of AI artifacts, ensuring auditability of agentic systems</p></li></ul><h2>Architectural Patterns for Production Systems</h2><h3>Multi-Agent Framework Design</h3><p>In my deployments of OpenAI&#8217;s framework, I&#8217;ve prioritized client-side execution for enterprises handling sensitive data like financial transactions. This approach minimizes server dependency but introduces tradeoffs: no persistent memory or shared state between agents. For static workflows like document processing, it works seamlessly. For dynamic supply chain scenarios requiring real-time ERP data integration, you&#8217;ll need supplemental vector databases.</p><p><strong>The choice between dynamic task routing and predefined workflows depends on risk tolerance. OpenAI&#8217;s delegation model allows agents to autonomously hand off subtasks (e.g., triage agent to research agent), while LangChain enforces rigid pipelines.</strong> Through trial and error, I&#8217;ve learned hybrid architectures work best: predefined modules for compliance-critical steps (audit trails, access controls) paired with dynamic routing for analyzing market trends.</p><p>When evaluating frameworks:</p><ul><li><p><strong>LangChain</strong> excels in custom chat interfaces but struggles with SOC2-compliant audit trails</p></li><li><p><strong>LlamaIndex</strong> dominates RAG implementations but lacks native tool execution for robotic process automation</p></li><li><p><strong>OpenAI&#8217;s framework</strong> sacrifices memory management for simplicity, requiring external systems for context-heavy workflows</p></li></ul><h2>Critical Subsystems</h2><h3>Governance Layer</h3><p>Production systems demand policy engines that surpass basic API key checks. I recommend implementing the following for healthcare:</p><ul><li><p><strong>Attribute-based access controls</strong> tied to Azure Active Directory roles</p></li><li><p><strong>Explainability pipelines</strong> generating timestamped audit trails with decision rationales, critical for HIPAA compliance</p></li><li><p><strong>Circuit breakers</strong> freezing agent pools during anomaly detection events (e.g., abnormal inventory spikes in supply chain agents)</p></li></ul><h3>Performance Optimization</h3><p> Scaling agentic systems requires intelligent resource allocation:</p><ul><li><p><strong>Agent pooling</strong> maintains hot standby instances for high-priority tasks like processing real-time sensor data</p></li><li><p><strong>Cold start mitigation</strong> uses synthetic data mimicking historical enterprise system interactions to pre-warm agents</p></li><li><p><strong>Cost-aware execution</strong> reduced LLM API costs 40% in my projects by routing basic interactions to Mistral-7B while reserving GPT-4 Turbo for complex scenario decomposition</p></li></ul><p>Always implement two-tier logging&#8212;structured JSON for systems-of-record and human-readable summaries for oversight teams.</p><h2>Implementation Roadmap</h2><h3>Phase 1: Agent Readiness Assessment</h3><p>Start by <strong>process mining</strong> enterprise workflows to identify automation candidates. In my work with healthcare clients, tools like Microsoft Power Automate&#8217;s process mining module uncovered 40% efficiency gains in patient data routing by analyzing EHR interaction patterns. Simultaneously, evaluate your <strong>API ecosystem maturity</strong>.</p><p><strong>Key activities</strong>:</p><ul><li><p>Map high-frequency, low-variance tasks (insurance claims processing, IT ticket routing)</p></li><li><p>Audit API endpoints for scalability, security, and compatibility with agentic systems</p></li><li><p>Conduct a <strong>sensitive data inventory</strong> to identify access control requirements</p></li></ul><h3>Phase 2: Pilot Design</h3><p>Select <strong>bounded problem spaces</strong> where agents can deliver quick wins without disrupting complex workflows. Salesforce&#8217;s BBBSA pilot succeeded by focusing on mentor matching, a contained use case with clear success metrics. Build <strong>observability pipelines</strong> from day one.</p><p><strong>Critical elements</strong>:</p><ul><li><p><strong>Define agent performance SLAs (e.g., 95% accuracy in supply chain demand forecasting)</strong></p></li><li><p>Implement human oversight loops for high-stakes decisions (medical diagnoses, financial approvals)</p></li><li><p>Use synthetic data to simulate edge cases in autonomous systems</p></li></ul><h3>Phase 3: Scaling Challenges</h3><p><strong>Agent-to-agent communication</strong> becomes the bottleneck at scale. <strong>Emergent behaviors</strong> in complex systems require constant vigilance.</p><p><strong>Security imperatives</strong>:</p><ul><li><p>Enforce least-privilege access controls across autonomous agents</p></li><li><p>Conduct adversarial testing for unintended consequences in AI-driven actions</p></li><li><p>Implement Zero Trust architecture for agent-to-enterprise system interactions</p></li></ul><div><hr></div><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!-kh9!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!-kh9!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!-kh9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png" width="1456" height="246" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:246,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:172293,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:&quot;&quot;,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!-kh9!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p>I also host an AI podcast and content series called &#8220;Pioneers.&#8221; This series takes you on an enthralling journey into the minds of AI visionaries, founders, and CEOs who are at the forefront of innovation through AI in their organizations.</p><p>To learn more, please visit <a href="https://www.multimodal.dev/blog">Pioneers</a> on Beehiiv.</p><div><hr></div><h2>Wrapping Up</h2><p>To harness agentic AI's potential, prioritize <strong>adaptive architectures</strong> over rigid systems&#8212;modular designs with hybrid memory (vector DBs + RAG) enable dynamic problem-solving across supply chains and IT operations. Implement <strong>governance-first development</strong>: bake policy engines and audit trails into agent frameworks early, as financial institutions do for real-time transaction monitoring.</p><p><strong>Key technical advice</strong>:</p><ul><li><p>Start with bounded pilots (e.g., SAP order management) using synthetic data to mitigate cold starts</p></li><li><p>Design multi-agent systems with circuit breakers and privilege tiers&#8212;healthcare orgs reduced patient data breaches 60% through Azure AD-integrated access controls</p></li><li><p>Treat agents as team members: retrain quarterly with fresh market trends and conduct performance reviews using CI/CD metrics</p></li></ul><p>The future belongs to <strong>human-agent symbiosis</strong>&#8212;where logistics bots negotiate rates and your team focuses on strategic innovation. As AWS implementations show, production-ready systems require obsessive instrumentation: two-tier logging (structured + human-readable) and CloudWatch dashboards for emergent behavior detection. Start small, but think in workflows, not tasks.</p><p>Until next week,<br>Ankur.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.ankursnewsletter.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">If you liked reading this post, consider subscribing for more AI content.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Key RAG Techniques: Benefits, Costs, Applications]]></title><description><![CDATA[Compare different Retrieval Augmented Generation (RAG) techniques including implementation methods, best practices, and challenges to overcome.]]></description><link>https://www.ankursnewsletter.com/p/key-rag-techniques-benefits-costs</link><guid isPermaLink="false">https://www.ankursnewsletter.com/p/key-rag-techniques-benefits-costs</guid><dc:creator><![CDATA[Ankur A. Patel]]></dc:creator><pubDate>Fri, 06 Jun 2025 19:36:24 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!bTjE!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fee537cc1-1249-444c-ac9b-06348504d465_1200x600.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!bTjE!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fee537cc1-1249-444c-ac9b-06348504d465_1200x600.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!bTjE!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fee537cc1-1249-444c-ac9b-06348504d465_1200x600.png 424w, https://substackcdn.com/image/fetch/$s_!bTjE!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fee537cc1-1249-444c-ac9b-06348504d465_1200x600.png 848w, https://substackcdn.com/image/fetch/$s_!bTjE!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fee537cc1-1249-444c-ac9b-06348504d465_1200x600.png 1272w, https://substackcdn.com/image/fetch/$s_!bTjE!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fee537cc1-1249-444c-ac9b-06348504d465_1200x600.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!bTjE!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fee537cc1-1249-444c-ac9b-06348504d465_1200x600.png" width="1200" height="600" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/ee537cc1-1249-444c-ac9b-06348504d465_1200x600.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:600,&quot;width&quot;:1200,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:378605,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.ankursnewsletter.com/i/165369315?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fee537cc1-1249-444c-ac9b-06348504d465_1200x600.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!bTjE!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fee537cc1-1249-444c-ac9b-06348504d465_1200x600.png 424w, https://substackcdn.com/image/fetch/$s_!bTjE!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fee537cc1-1249-444c-ac9b-06348504d465_1200x600.png 848w, https://substackcdn.com/image/fetch/$s_!bTjE!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fee537cc1-1249-444c-ac9b-06348504d465_1200x600.png 1272w, https://substackcdn.com/image/fetch/$s_!bTjE!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fee537cc1-1249-444c-ac9b-06348504d465_1200x600.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.ankursnewsletter.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Subscribe today to join our community of 2000 AI enthusiasts and business leaders for more insider takes and practical advice on implementing AI in business.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>Let&#8217;s revisit one of our most popular posts of 2024 around RAG and its implementation. As enterprises today continue to navigate the agentic AI landscape, this topic is more timely than ever. </p><h2><strong>Key Takeaways</strong></h2><p>1. Retrieval Augmented Generation (RAG) is essential for effective <strong>enterprise AI implementation</strong>, enhancing information retrieval and overcoming limitations of traditional large language models.</p><p>2. Four main RAG techniques include <strong>traditional document-based, vector database, hybrid, and multimodal approaches</strong>, each with unique strengths and applications.</p><p>3. Choosing the right <strong>RAG method</strong> depends on an enterprise's specific needs, data types, and computational resources.</p><p>4. Successful RAG implementation requires careful consideration of <strong>data quality, security, scalability, and integration</strong> with existing systems.</p><p>5. Best practices for RAG deployment include <strong>prioritizing data management, investing in security, embracing an API-first approach</strong>, and starting with pilot projects before full-scale implementation.</p><div><hr></div><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!R9ua!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!R9ua!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!R9ua!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png" width="1456" height="246" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/bea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:246,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:273651,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:&quot;&quot;,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!R9ua!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1456w" sizes="100vw" loading="lazy" fetchpriority="high"></picture><div></div></div></a></figure></div><p>In 2023, I started Multimodal, a Generative AI company that helps organizations automate complex, knowledge-based workflows using AI Agents. Check it out <a href="https://www.multimodal.dev/">here</a>.</p><div><hr></div><p>Retrieval Augmented Generation (RAG) is key to enterprises unlocking AI implementation and making their workflows smoother and more efficient. Generative AI is truly the future of work, but for it to be effective for complex knowledge work, techniques like RAG are crucial.</p><p>Let&#8217;s dive into some of the <a href="https://www.ankursnewsletter.com/p/making-unstructured-data-ai-ready?utm_source=publication-search">most effective ways to do RAG</a>, and how you should think about it if you&#8217;re trying to build AI systems for enterprises.</p><h2><strong>Why enterprise AI needs RAG</strong></h2><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!7cnl!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F31a8196c-e946-4db0-961e-1844c7a1509b_800x454.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!7cnl!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F31a8196c-e946-4db0-961e-1844c7a1509b_800x454.png 424w, https://substackcdn.com/image/fetch/$s_!7cnl!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F31a8196c-e946-4db0-961e-1844c7a1509b_800x454.png 848w, https://substackcdn.com/image/fetch/$s_!7cnl!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F31a8196c-e946-4db0-961e-1844c7a1509b_800x454.png 1272w, https://substackcdn.com/image/fetch/$s_!7cnl!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F31a8196c-e946-4db0-961e-1844c7a1509b_800x454.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!7cnl!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F31a8196c-e946-4db0-961e-1844c7a1509b_800x454.png" width="800" height="454" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/31a8196c-e946-4db0-961e-1844c7a1509b_800x454.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:454,&quot;width&quot;:800,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!7cnl!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F31a8196c-e946-4db0-961e-1844c7a1509b_800x454.png 424w, https://substackcdn.com/image/fetch/$s_!7cnl!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F31a8196c-e946-4db0-961e-1844c7a1509b_800x454.png 848w, https://substackcdn.com/image/fetch/$s_!7cnl!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F31a8196c-e946-4db0-961e-1844c7a1509b_800x454.png 1272w, https://substackcdn.com/image/fetch/$s_!7cnl!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F31a8196c-e946-4db0-961e-1844c7a1509b_800x454.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><a href="https://bradsol.com/wp-content/uploads/2024/04/image-1-RAG-e1713883898899.jpg">Source</a></figcaption></figure></div><p>Large language models are excellent at doing complex tasks really fast. But because of their huge training database and complicated architecture, they&#8217;re not very relevant for enterprise applications without some tweaking. Here&#8217;s why I recommend every enterprise AI builder to engage with RAG:</p><p>1. <strong>Enhanced information retrieval</strong>: RAG combines large language models (LLMs) with external knowledge sources, providing accurate and contextually relevant responses to user queries.</p><p>2. <strong>Overcoming LLM limitations</strong>: Traditional LLMs often struggle with real-time or domain-specific inquiries. RAG addresses this by integrating up-to-date information.</p><p>3. <strong>Streamlined workflows</strong>: In finance and insurance, RAG systems can integrate structured and unstructured data sources, improving decision-making processes.</p><h2><strong>RAG implementation methods</strong></h2><p>Depending on the way you&#8217;re going to use AI for your enterprise, the method of retrieval augmented generation you choose will also differ. Here are some typical RAG techniques enterprises use:</p><h3><strong>1. Traditional document-based RAG</strong></h3><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!J9kx!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F350466ec-002a-46ca-8af7-797d607f7dbc_1600x943.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!J9kx!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F350466ec-002a-46ca-8af7-797d607f7dbc_1600x943.png 424w, https://substackcdn.com/image/fetch/$s_!J9kx!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F350466ec-002a-46ca-8af7-797d607f7dbc_1600x943.png 848w, https://substackcdn.com/image/fetch/$s_!J9kx!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F350466ec-002a-46ca-8af7-797d607f7dbc_1600x943.png 1272w, https://substackcdn.com/image/fetch/$s_!J9kx!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F350466ec-002a-46ca-8af7-797d607f7dbc_1600x943.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!J9kx!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F350466ec-002a-46ca-8af7-797d607f7dbc_1600x943.png" width="1456" height="858" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/350466ec-002a-46ca-8af7-797d607f7dbc_1600x943.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:858,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!J9kx!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F350466ec-002a-46ca-8af7-797d607f7dbc_1600x943.png 424w, https://substackcdn.com/image/fetch/$s_!J9kx!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F350466ec-002a-46ca-8af7-797d607f7dbc_1600x943.png 848w, https://substackcdn.com/image/fetch/$s_!J9kx!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F350466ec-002a-46ca-8af7-797d607f7dbc_1600x943.png 1272w, https://substackcdn.com/image/fetch/$s_!J9kx!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F350466ec-002a-46ca-8af7-797d607f7dbc_1600x943.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><a href="https://www.promptingguide.ai/_next/image?url=%2F_next%2Fstatic%2Fmedia%2Frag-framework.81dc2cdc.png&amp;w=3840&amp;q=75">Source</a></figcaption></figure></div><p>This method forms the foundation of many <a href="https://www.multimodal.dev/post/how-to-build-a-rag-pipeline">RAG systems</a>, especially in text-heavy industries. Here's what you need to know:</p><h4><strong>How it works</strong></h4><ul><li><p><strong>Document ingestion</strong>: The system ingests and processes documents from various sources, such as research reports, customer service guides, and web pages.</p></li><li><p><strong>Indexing</strong>: Documents are indexed for efficient retrieval, often using keyword search or basic query techniques.</p></li><li><p><strong>Retrieval</strong>: When a user submits a query, the system searches the indexed documents for relevant information.</p></li><li><p><strong>Augmentation</strong>: Retrieved passages are used to augment the prompt given to the large language model (LLM).</p></li><li><p><strong>Generation</strong>: The LLM generates a response based on the augmented prompt.</p></li></ul><h4><strong>Pros</strong></h4><p>1. <strong>Simple implementation</strong>: Leverages existing document management systems, making it easier to adopt.</p><p>2. <strong>Familiarity</strong>: Works well with traditional information retrieval methods that many organizations already use.</p><p>3. <strong>Domain specificity</strong>: Excels in industries with extensive textual documentation, like insurance and finance.</p><h4><strong>Cons</strong></h4><p>1. <strong>Limited data types</strong>: Primarily works with unstructured text data, potentially missing insights from other data formats.</p><p>2. <strong>Real-time challenges</strong>: May struggle to incorporate up-to-date information if document repositories aren't frequently updated.</p><p>3. <strong>Scalability concerns</strong>: As the document library grows, retrieval speed and accuracy can be impacted without optimized search algorithms.</p><h4><strong>Enterprise applications</strong></h4><ul><li><p><strong>Policy document retrieval</strong>: Insurance companies can use this method to quickly access relevant policy information when addressing customer queries.</p></li><li><p><strong>Financial report analysis</strong>: Investment firms can leverage RAG to extract key insights from vast repositories of financial reports and market analyses.</p></li></ul><h3><strong>2. Vector database RAG</strong></h3><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!CKV1!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F627af0fd-b8d3-4e6b-8474-9c88b03ea8c1_948x660.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!CKV1!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F627af0fd-b8d3-4e6b-8474-9c88b03ea8c1_948x660.png 424w, https://substackcdn.com/image/fetch/$s_!CKV1!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F627af0fd-b8d3-4e6b-8474-9c88b03ea8c1_948x660.png 848w, https://substackcdn.com/image/fetch/$s_!CKV1!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F627af0fd-b8d3-4e6b-8474-9c88b03ea8c1_948x660.png 1272w, https://substackcdn.com/image/fetch/$s_!CKV1!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F627af0fd-b8d3-4e6b-8474-9c88b03ea8c1_948x660.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!CKV1!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F627af0fd-b8d3-4e6b-8474-9c88b03ea8c1_948x660.png" width="948" height="660" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/627af0fd-b8d3-4e6b-8474-9c88b03ea8c1_948x660.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:660,&quot;width&quot;:948,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!CKV1!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F627af0fd-b8d3-4e6b-8474-9c88b03ea8c1_948x660.png 424w, https://substackcdn.com/image/fetch/$s_!CKV1!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F627af0fd-b8d3-4e6b-8474-9c88b03ea8c1_948x660.png 848w, https://substackcdn.com/image/fetch/$s_!CKV1!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F627af0fd-b8d3-4e6b-8474-9c88b03ea8c1_948x660.png 1272w, https://substackcdn.com/image/fetch/$s_!CKV1!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F627af0fd-b8d3-4e6b-8474-9c88b03ea8c1_948x660.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><a href="https://miro.medium.com/v2/resize:fit:948/1*t7wvIqESWr3y2XloUdvPpQ.png">Source</a></figcaption></figure></div><h4><strong>What are vector databases?</strong></h4><p><a href="https://www.cloudflare.com/learning/ai/what-is-vector-database/">Vector databases </a>are specialized systems designed to store and efficiently search high-dimensional numerical representations of data, known as embeddings. These embeddings capture the semantic meaning of information, allowing for more nuanced and context-aware retrieval compared to traditional keyword search methods.</p><h4><strong>How RAG leverages vector search capabilities</strong></h4><p>1. <strong>Embedding generation</strong>: When a user submits a query, it's converted into a vector embedding using an embedding model.</p><p>2. <strong>Similarity search</strong>: The system searches the vector database for the most similar embeddings to the query vector.</p><p>3. <strong>Retrieval</strong>: Relevant documents or data points associated with the similar embeddings are retrieved.</p><p>4. <strong>Augmentation</strong>: The retrieved information is used to augment the prompt sent to the large language model (LLM).</p><p>5. <strong>Response generation</strong>: The LLM generates a response based on the augmented prompt, combining its training data with the retrieved external knowledge.</p><h4><strong>Pros</strong></h4><p>1. <strong>Efficient semantic search</strong>: Vector databases excel at finding semantically relevant passages, even when exact keyword matches aren't present.</p><p>2. <strong>Handles diverse data types</strong>: From text and images to audio and video, vector databases can process and search various data formats.</p><p>3. <strong>Scalability</strong>: Designed to handle large datasets, making them suitable for enterprises with vast knowledge bases.</p><h4><strong>Cons</strong></h4><p>1. <strong>Initial setup complexity</strong>: Implementing this kind of RAG system requires careful planning and expertise in data transformation and embedding generation.</p><p>2. <strong>Periodic reindexing</strong>: To maintain optimal performance, the database may need regular updates and reindexing, especially when dealing with rapidly changing information.</p><p>3. <strong>Computational requirements</strong>: Vector search operations can be computationally intensive, potentially leading to higher infrastructure costs.</p><h4><strong>Enterprise applications</strong></h4><p>1. <strong>Real-time market data analysis in finance</strong>: Vector databases enable financial institutions to quickly process and analyze vast amounts of market data, news, and research reports. This allows for more accurate and timely investment decisions.</p><p>2. <strong>Customer profile matching in insurance</strong>: Insurance companies can use vector database RAG to match customer profiles with the most relevant policies or risk assessments.</p><p>Remember, while vector databases offer powerful capabilities, their successful implementation requires careful consideration of your specific use case, data types, and computational resources.</p><h3><strong>3. Hybrid RAG (combining structured and unstructured data)</strong></h3><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!PUHD!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd730db46-2500-43a2-8a56-dc68510f25be_1600x835.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!PUHD!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd730db46-2500-43a2-8a56-dc68510f25be_1600x835.png 424w, https://substackcdn.com/image/fetch/$s_!PUHD!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd730db46-2500-43a2-8a56-dc68510f25be_1600x835.png 848w, https://substackcdn.com/image/fetch/$s_!PUHD!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd730db46-2500-43a2-8a56-dc68510f25be_1600x835.png 1272w, https://substackcdn.com/image/fetch/$s_!PUHD!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd730db46-2500-43a2-8a56-dc68510f25be_1600x835.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!PUHD!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd730db46-2500-43a2-8a56-dc68510f25be_1600x835.png" width="1456" height="760" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d730db46-2500-43a2-8a56-dc68510f25be_1600x835.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:760,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!PUHD!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd730db46-2500-43a2-8a56-dc68510f25be_1600x835.png 424w, https://substackcdn.com/image/fetch/$s_!PUHD!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd730db46-2500-43a2-8a56-dc68510f25be_1600x835.png 848w, https://substackcdn.com/image/fetch/$s_!PUHD!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd730db46-2500-43a2-8a56-dc68510f25be_1600x835.png 1272w, https://substackcdn.com/image/fetch/$s_!PUHD!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd730db46-2500-43a2-8a56-dc68510f25be_1600x835.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><a href="https://adasci.org/wp-content/uploads/2024/09/Screenshot-2024-09-12-at-4.51.03%E2%80%AFPM.png">Source</a></figcaption></figure></div><p>This innovative approach to Retrieval Augmented Generation is transforming how businesses handle knowledge-intensive tasks and process complex user queries.</p><h4><strong>How it works</strong></h4><p>Hybrid retrieval augmented generation integrates traditional databases with unstructured data sources, creating a comprehensive knowledge ecosystem. Here's a breakdown of the process:</p><p>1. <strong>Data ingestion</strong>: The system ingests both structured data (e.g., from <a href="https://www.solarwinds.com/resources/it-glossary/sql-database">SQL databases</a>) and unstructured data (e.g., documents, web pages, and customer service guides).</p><p>2. <strong>Unified indexing</strong>: A sophisticated indexing system creates a bridge between structured and unstructured data, often using vector databases for efficient retrieval.</p><p>3. <strong>Query processing</strong>: When a user submits a query, the system searches both structured and unstructured data sources simultaneously.</p><p>4. <strong>Relevance ranking</strong>: Using advanced algorithms, the system ranks the most relevant information from both data types.</p><p>5. <strong>Augmented prompt creation</strong>: The retrieved information is used to create an augmented prompt for the large language model (LLM).</p><p>6. <strong>Response generation</strong>: The LLM generates a comprehensive response based on the augmented prompt, leveraging both structured and unstructured data insights.</p><h4><strong>Pros</strong></h4><p>1. <strong>Comprehensive data utilization</strong>: Hybrid RAG taps into the full spectrum of enterprise data, from transactional records to unstructured text, providing a 360-degree view of information.</p><p>2. <strong>Enhanced accuracy</strong>: By combining structured data's precision with unstructured data's context, Hybrid retrieval augmented generation produces more accurate and nuanced responses.</p><p>3. <strong>Adaptability</strong>: This approach can be tailored to various enterprise data ecosystems, making it versatile across different industries and use cases.</p><h4><strong>Cons</strong></h4><p>1. <strong>Complex implementation</strong>: Integrating disparate data sources requires sophisticated data integration strategies and expertise.</p><p>2. <strong>Potential latency</strong>: Querying multiple data sources simultaneously may lead to increased response times, especially for complex queries.</p><p>3. <strong>Data governance challenges</strong>: Handling both structured and unstructured data raises concerns about data privacy and security, particularly when dealing with sensitive customer information.</p><h4><strong>Enterprise applications</strong></h4><p>1. <strong>Risk assessment in insurance</strong>: <strong>Hybrid RAG excels in combining structured policy data with unstructured external factors like weather reports, social media trends, and news articles.</strong> This comprehensive approach enables insurers to make more accurate risk assessments and offer personalized policies.</p><p>2. <strong>Fraud detection in finance</strong>: By analyzing structured transaction data alongside unstructured customer information (e.g., support tickets, social media activity), financial institutions can identify suspicious patterns more effectively.</p><p>3. <strong>Customer service enhancement</strong>: Hybrid retrieval augmented generation can provide customer service representatives with a holistic view of customer data, combining structured account information with unstructured interaction history. This enables more personalized and efficient support.</p><p>Hybrid RAG bridges the gap between structured and unstructured data, enabling organizations to unlock deeper insights, make more informed decisions, and provide superior services to their customers.</p><h3><strong>4. Multimodal RAG</strong></h3><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!6CJZ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcc3b86f7-e994-4357-bc02-2bee6272e80e_982x733.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!6CJZ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcc3b86f7-e994-4357-bc02-2bee6272e80e_982x733.png 424w, https://substackcdn.com/image/fetch/$s_!6CJZ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcc3b86f7-e994-4357-bc02-2bee6272e80e_982x733.png 848w, https://substackcdn.com/image/fetch/$s_!6CJZ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcc3b86f7-e994-4357-bc02-2bee6272e80e_982x733.png 1272w, https://substackcdn.com/image/fetch/$s_!6CJZ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcc3b86f7-e994-4357-bc02-2bee6272e80e_982x733.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!6CJZ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcc3b86f7-e994-4357-bc02-2bee6272e80e_982x733.png" width="982" height="733" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/cc3b86f7-e994-4357-bc02-2bee6272e80e_982x733.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:733,&quot;width&quot;:982,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!6CJZ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcc3b86f7-e994-4357-bc02-2bee6272e80e_982x733.png 424w, https://substackcdn.com/image/fetch/$s_!6CJZ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcc3b86f7-e994-4357-bc02-2bee6272e80e_982x733.png 848w, https://substackcdn.com/image/fetch/$s_!6CJZ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcc3b86f7-e994-4357-bc02-2bee6272e80e_982x733.png 1272w, https://substackcdn.com/image/fetch/$s_!6CJZ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcc3b86f7-e994-4357-bc02-2bee6272e80e_982x733.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><a href="https://miro.medium.com/v2/resize:fit:982/1*4CaYSUEN9Z51bFgnFLXFOg.png">Source</a></figcaption></figure></div><p>This innovative approach extends the capabilities of traditional RAG systems by incorporating diverse data types, including text, images, and audio.</p><h4><strong>How it works</strong></h4><p>Multimodal RAG integrates various data formats into a unified retrieval and generation system:</p><p>1. <strong>Data ingestion</strong>: The system ingests diverse data types, including text documents, images, audio files, and even video content.</p><p>2. <strong>Multimodal embedding</strong>: Advanced embedding models convert different data types into a common vector space, allowing for unified indexing and retrieval.</p><p>3. <strong>Cross-modal retrieval</strong>: When a user submits a query (which can be text, image, or audio), the system searches across all data types to find relevant information.</p><p>4. <strong>Relevance ranking</strong>: Sophisticated algorithms rank the most relevant information across modalities.</p><p>5. <strong>Multimodal prompt creation</strong>: The retrieved information, regardless of its original format, is used to create an augmented prompt for the large language model (LLM).</p><p>6. <strong>Response generation</strong>: The LLM generates a comprehensive response, potentially incorporating insights from various data types.</p><h4><strong>Pros</strong></h4><p>1. <strong>Comprehensive data utilization</strong>: Multimodal RAG taps into a broader spectrum of enterprise data, providing a more holistic view of information.</p><p>2.<strong> Enhanced context understanding</strong>: By incorporating visual and audio data, the system can capture nuances that text alone might miss.</p><p>3. <strong>Improved accuracy in complex scenarios</strong>: For tasks requiring multi-faceted analysis, such as insurance claim assessment, multimodal RAG can provide more accurate and nuanced responses.</p><h4><strong>Cons</strong></h4><p>1.<strong> Increased complexity</strong>: Handling multiple data types requires more sophisticated algorithms and infrastructure, potentially increasing implementation challenges.</p><p>2. <strong>Higher computational requirements</strong>: Processing and analyzing diverse data formats can be computationally intensive, potentially leading to increased costs.</p><p>3. <strong>Data quality challenges</strong>: Ensuring consistent quality across different data types can be more challenging than with text-only systems.</p><h4><strong>Enterprise applications</strong></h4><p>1. <strong>Visual damage analysis</strong>: In auto insurance claims, the system can analyze photos of vehicle damage alongside textual claim descriptions, providing more accurate assessments.</p><p>2. <strong>Audio statement processing</strong>: For personal injury claims, the system can transcribe and analyze audio statements from claimants, witnesses, and experts, correlating this information with written reports and visual evidence.</p><p>3. <strong>Document verification</strong>: The system can cross-reference handwritten claim forms with typed documents and database records, flagging discrepancies more effectively.</p><p>4. <strong>Fraud detection</strong>: By analyzing patterns across text, image, and audio data, the system can identify potential fraudulent claims more accurately than text-only systems.</p><p>5. <strong>Real-time claim estimation</strong>: Adjusters in the field can submit photos and voice notes, receiving instant preliminary assessments based on historical claim data and visual analysis.</p><p>Multimodal RAG represents a significant leap forward in enterprise AI capabilities, particularly for industries like insurance that deal with diverse data types.</p><h2><strong>Comparative analysis</strong></h2><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!uLMY!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd4c33fd-0e00-401d-b5e6-96b6361d979d_1493x492.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!uLMY!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd4c33fd-0e00-401d-b5e6-96b6361d979d_1493x492.png 424w, https://substackcdn.com/image/fetch/$s_!uLMY!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd4c33fd-0e00-401d-b5e6-96b6361d979d_1493x492.png 848w, https://substackcdn.com/image/fetch/$s_!uLMY!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd4c33fd-0e00-401d-b5e6-96b6361d979d_1493x492.png 1272w, https://substackcdn.com/image/fetch/$s_!uLMY!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd4c33fd-0e00-401d-b5e6-96b6361d979d_1493x492.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!uLMY!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd4c33fd-0e00-401d-b5e6-96b6361d979d_1493x492.png" width="1456" height="480" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/dd4c33fd-0e00-401d-b5e6-96b6361d979d_1493x492.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:480,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;Comparative analysis of different RAG methods&quot;,&quot;title&quot;:&quot;Comparative analysis of different RAG methods&quot;,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Comparative analysis of different RAG methods" title="Comparative analysis of different RAG methods" srcset="https://substackcdn.com/image/fetch/$s_!uLMY!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd4c33fd-0e00-401d-b5e6-96b6361d979d_1493x492.png 424w, https://substackcdn.com/image/fetch/$s_!uLMY!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd4c33fd-0e00-401d-b5e6-96b6361d979d_1493x492.png 848w, https://substackcdn.com/image/fetch/$s_!uLMY!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd4c33fd-0e00-401d-b5e6-96b6361d979d_1493x492.png 1272w, https://substackcdn.com/image/fetch/$s_!uLMY!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdd4c33fd-0e00-401d-b5e6-96b6361d979d_1493x492.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Comparative analysis of different RAG methods</figcaption></figure></div><p>When evaluating <a href="https://www.multimodal.dev/post/rag-pipeline-diagram">RAG implementation</a> methods for enterprise use, it's essential to consider various factors that can impact performance, scalability, and integration capabilities. Let's break down how traditional document-based RAG, vector database RAG, hybrid RAG, and multimodal RAG compare across key metrics:</p><ul><li><p><strong>Implementation complexity</strong>: Traditional document-based RAG is relatively straightforward to implement, especially for organizations with existing document management systems. <strong>On the other hand, multimodal RAG requires sophisticated algorithms and infrastructure to handle diverse data types, making it the most complex to implement.</strong></p></li></ul><ul><li><p><strong>Scalability</strong>: Vector database RAG, hybrid RAG, and multimodal RAG all offer high scalability, thanks to their efficient indexing and retrieval mechanisms. Traditional document-based RAG may face challenges with very large document sets.</p></li></ul><ul><li><p><strong>Data type handling</strong>: Multimodal retrieval augmented generation excels here, capable of processing text, images, and audio. Hybrid RAG follows closely, adept at handling both structured and unstructured data. Vector database RAG is primarily focused on text and structured data, while traditional document-based RAG is limited mostly to text.</p></li></ul><ul><li><p><strong>Real-time capabilities</strong>: Vector database retrieval augmented generation, hybrid RAG, and multimodal RAG all offer strong real-time capabilities, crucial for applications like real-time market data analysis in finance or instant claim assessment in insurance. Traditional document-based RAG may lag slightly in this area.</p></li></ul><ul><li><p><strong>Integration with existing systems</strong>: Traditional document-based RAG often integrates seamlessly with existing document management systems. <strong>Hybrid RAG also tends to integrate well, given its ability to work with both structured and unstructured data sources.</strong> Vector database and multimodal RAG may require more adaptation of existing systems.</p></li></ul><p>When choosing between these techniques, my first step is to analyze the nature of the enterprise and the key workflows it wants a solution for. For example, a financial institution dealing primarily with textual data might find vector database RAG sufficient, while an insurance company handling diverse data types for claim assessment might benefit more from multimodal RAG.</p><h2><strong>Best practices for enterprise RAG implementation</strong></h2><h3><strong>1. Choosing appropriate knowledge sources</strong></h3><p>- Identify relevant internal and external data sources</p><p>- Prioritize high-quality, up-to-date information</p><p>- Consider domain-specific knowledge bases</p><h3><strong>2. Data preparation and management</strong></h3><p>- Clean and structure data for optimal retrieval</p><p>- Implement robust data governance policies</p><p>- Regularly update and maintain knowledge libraries</p><h3><strong>3. Fine-tuning strategies</strong></h3><p>- Adapt large language models to your specific domain</p><p>- Use domain-specific training data for better performance</p><p>- Continuously refine models based on user feedback</p><h3><strong>4. Updating and maintaining knowledge libraries</strong></h3><p>- Establish processes for regular content updates</p><p>- Implement version control for knowledge bases</p><p>- Monitor and evaluate retrieval performance</p><h2><strong>Challenges and considerations</strong></h2><p>While RAG offers significant benefits, enterprises must navigate several challenges:</p><h3><strong>1. Scalability and performance optimization</strong></h3><p>- Design systems to handle increasing data volumes</p><p>- Optimize retrieval algorithms for faster response times</p><p>- Balance computational costs with performance requirements</p><h3><strong>2. Data privacy and security</strong></h3><p>- Implement robust encryption for sensitive data</p><p>- Ensure compliance with data protection regulations</p><p>- Manage access controls for personally identifiable information</p><h3><strong>3. Integration with existing enterprise systems</strong></h3><p>- Develop APIs for seamless integration</p><p>- Ensure compatibility with legacy systems</p><p>- Address potential conflicts with existing workflows</p><div><hr></div><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!-kh9!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!-kh9!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!-kh9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png" width="1456" height="246" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:246,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:172293,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:&quot;&quot;,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!-kh9!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p>I also host an AI podcast and content series called &#8220;Pioneers.&#8221; This series takes you on an enthralling journey into the minds of AI visionaries, founders, and CEOs who are at the forefront of innovation through AI in their organizations.</p><p>To learn more, please visit <a href="https://www.multimodal.dev/blog">Pioneers</a> on Beehiiv.</p><div><hr></div><h2><strong>Wrapping up</strong></h2><p>If you&#8217;re looking to implement RAG in your enterprise, here&#8217;s what I recommend you keep in mind:</p><p>1. <strong>Prioritize data quality and management</strong>: Ensure your data sources are accurate, up-to-date, and well-structured. Implement regular data cleaning and updating processes to maintain the integrity of your knowledge base.</p><p>2. <strong>Invest in robust security measures</strong>: Implement strong encryption, access controls, and data masking techniques to protect sensitive information. Regularly audit your RAG system for potential vulnerabilities.</p><p>3. <strong>Focus on scalability</strong>: Design your RAG architecture with growth in mind. Consider cloud-based solutions or hybrid approaches that can easily scale with your needs.</p><p>4. <strong>Embrace an API-first approach</strong>: This ensures seamless integration with existing systems and allows for greater flexibility as your needs evolve.</p><p>5. <strong>Implement fine-tuning strategies</strong>: Adapt your RAG model to your specific domain for improved performance and relevance.</p><p>6. <strong>Address ethical considerations</strong>: Be mindful of potential biases in your data and model outputs. Implement safeguards to ensure fair and ethical use of the technology.</p><p>7. <strong>Invest in employee training</strong>: Ensure your team has the necessary skills to manage and maintain your RAG system effectively.</p><p>8. <strong>Start small and iterate</strong>: Begin with pilot projects to gain experience and refine your approach before full-scale implementation.</p><p>9.<strong> Prioritize user experience</strong>: Design your RAG-powered applications with a focus on user-friendliness and intuitive interactions.</p><p>10. <strong>Stay compliant</strong>: Ensure your RAG implementation adheres to relevant data privacy regulations and industry standards.</p><p>Next week, we will explore more about enterprise AI and agentic automation.</p><p>Until then,<br>Ankur</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.ankursnewsletter.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">If you liked reading this post, consider subscribing for more AI content.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Gemini Vs Claude: What’s Better at Coding?]]></title><description><![CDATA[Compare Google Gemini vs Anthropic&#8217;s Claude for coding, multimodal integration, agentic automation, benchmark results, and real-world use cases.]]></description><link>https://www.ankursnewsletter.com/p/gemini-vs-claude-whats-better-at</link><guid isPermaLink="false">https://www.ankursnewsletter.com/p/gemini-vs-claude-whats-better-at</guid><dc:creator><![CDATA[Ankur A. Patel]]></dc:creator><pubDate>Thu, 29 May 2025 18:47:10 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!S57M!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b0997db-3b8e-4b0c-9182-330e95b862fa_1200x600.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!S57M!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b0997db-3b8e-4b0c-9182-330e95b862fa_1200x600.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!S57M!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b0997db-3b8e-4b0c-9182-330e95b862fa_1200x600.png 424w, https://substackcdn.com/image/fetch/$s_!S57M!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b0997db-3b8e-4b0c-9182-330e95b862fa_1200x600.png 848w, https://substackcdn.com/image/fetch/$s_!S57M!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b0997db-3b8e-4b0c-9182-330e95b862fa_1200x600.png 1272w, https://substackcdn.com/image/fetch/$s_!S57M!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b0997db-3b8e-4b0c-9182-330e95b862fa_1200x600.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!S57M!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b0997db-3b8e-4b0c-9182-330e95b862fa_1200x600.png" width="1200" height="600" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/7b0997db-3b8e-4b0c-9182-330e95b862fa_1200x600.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:600,&quot;width&quot;:1200,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:375756,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.ankursnewsletter.com/i/164734130?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b0997db-3b8e-4b0c-9182-330e95b862fa_1200x600.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!S57M!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b0997db-3b8e-4b0c-9182-330e95b862fa_1200x600.png 424w, https://substackcdn.com/image/fetch/$s_!S57M!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b0997db-3b8e-4b0c-9182-330e95b862fa_1200x600.png 848w, https://substackcdn.com/image/fetch/$s_!S57M!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b0997db-3b8e-4b0c-9182-330e95b862fa_1200x600.png 1272w, https://substackcdn.com/image/fetch/$s_!S57M!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b0997db-3b8e-4b0c-9182-330e95b862fa_1200x600.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.ankursnewsletter.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Subscribe today to join our community of 2000 AI enthusiasts and business leaders for more insider takes and practical advice on implementing AI in business.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h2>Key Takeaways</h2><ol><li><p><strong>Gemini dominates multimodal coding</strong> (e.g., generating Tetris clones with dark themes) via seamless Google ecosystem integration, while <strong>Claude excels in marathon agentic workflows</strong> (e.g., porting chess games between languages) using million-token context retention.</p></li><li><p><strong>IDE warriors</strong> prefer Gemini for real-time VS Code/JetBrains autocompletion, whereas <strong>terminal purists</strong> choose Claude for one-shot commands handling Git commits and database migrations.</p></li><li><p><strong>Benchmark divergences</strong> show Gemini leading in HumanEval (85.3%) for rapid prototyping, while Claude rules SWE-bench (72.5%) for complex codebase refactors.</p></li><li><p><strong>Enterprise teams</strong> lean toward Gemini for SOC 2 compliance in financial apps, while <strong>privacy-focused startups</strong> adopt Claude for HIPAA-safe healthcare data pipelines.</p></li><li><p>The <strong>AI coding future</strong> lies in hybrid workflows: Gemini for Vertex AI data analysis &#8594; Claude for terminal-based physics engine debugging.</p></li></ol><div><hr></div><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!R9ua!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!R9ua!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!R9ua!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png" width="1456" height="246" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/bea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:246,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:273651,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:&quot;&quot;,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!R9ua!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1456w" sizes="100vw" loading="lazy" fetchpriority="high"></picture><div></div></div></a></figure></div><p>In 2023, I started Multimodal, a Generative AI company that helps organizations automate complex, knowledge-based workflows using AI Agents. Check it out <a href="https://www.multimodal.dev/">here</a>.</p><div><hr></div><p>The gemini vs claude debate underscores the AI boom&#8217;s impact on software development, where coding models now handle complex coding tasks.</p><p>Google Gemini leverages multimodal capabilities and seamless integration with Google Cloud&#8217;s Vertex AI, while Claude 3.7 Sonnet excels in logical reasoning, tackling chess game prompts or debugging real Mario game physics using its million-input-token context window.</p><p>In this article, I&#8217;ll compare the latest versions of both models. This comparison evaluates technical prowess (code generation, parallel test time compute), developer experience (user-friendly interfaces, Google AI Studio), and real-world coding applications.</p><h2><strong>Background: The Models and Their Evolution</strong></h2><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!XFOu!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd02236b7-510d-4db6-abcb-79438787b9ab_1311x581.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!XFOu!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd02236b7-510d-4db6-abcb-79438787b9ab_1311x581.png 424w, https://substackcdn.com/image/fetch/$s_!XFOu!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd02236b7-510d-4db6-abcb-79438787b9ab_1311x581.png 848w, https://substackcdn.com/image/fetch/$s_!XFOu!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd02236b7-510d-4db6-abcb-79438787b9ab_1311x581.png 1272w, https://substackcdn.com/image/fetch/$s_!XFOu!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd02236b7-510d-4db6-abcb-79438787b9ab_1311x581.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!XFOu!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd02236b7-510d-4db6-abcb-79438787b9ab_1311x581.png" width="1311" height="581" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d02236b7-510d-4db6-abcb-79438787b9ab_1311x581.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:581,&quot;width&quot;:1311,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:100337,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.ankursnewsletter.com/i/164734130?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd02236b7-510d-4db6-abcb-79438787b9ab_1311x581.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!XFOu!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd02236b7-510d-4db6-abcb-79438787b9ab_1311x581.png 424w, https://substackcdn.com/image/fetch/$s_!XFOu!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd02236b7-510d-4db6-abcb-79438787b9ab_1311x581.png 848w, https://substackcdn.com/image/fetch/$s_!XFOu!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd02236b7-510d-4db6-abcb-79438787b9ab_1311x581.png 1272w, https://substackcdn.com/image/fetch/$s_!XFOu!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd02236b7-510d-4db6-abcb-79438787b9ab_1311x581.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h2><strong>Gemini</strong></h2><ul><li><p><strong>Developer</strong>: Google DeepMind</p></li><li><p><strong>Core innovation</strong>: Natively multimodal architecture trained on text, code, images, and audio from inception, enabling seamless handling of tasks like generating a fully working chess game with implemented background theme music.</p></li><li><p><strong>Latest iteration</strong>: Gemini 2.5 Pro powers Gemini Code Assist, excelling at complex coding tasks such as creating a production-level chess game or debugging real Mario game physics using its million-input-token context window.</p></li><li><p><strong>Integration</strong>: Tightly coupled with Google Cloud's Vertex AI and Google AI Studio, offering a user-friendly interface for tasks ranging from Python script generation to data analysis.<br><br><strong>Strengths</strong>:</p><ul><li><p>Multimodal capabilities for cross-modal reasoning (e.g., interpreting dark theme UI mockups into functional code).</p></li><li><p>Seamless integration with Google Apps and GitHub Actions for CI/CD pipelines.</p></li></ul></li></ul><h2><strong>Anthropic Claude</strong></h2><ul><li><p><strong>Developer</strong>: Anthropic, building on Claude Shannon's information theory principles.</p></li><li><p><strong>Evolution</strong>: From Claude 3.7 Sonnet to Opus 4 and Sonnet 4, optimized for agentic workflows that handle complex problem solving across thousands of steps (e.g., 24-hour Pok&#233;mon Red guide generation).</p></li><li><p><strong>Coding specialization</strong>: Claude Code agent autonomously tackles tasks like refactoring legacy systems or implementing high score persistence via local storage, outperforming the previous best model by 3x in sustained task duration.</p><p></p><p><strong>Differentiators</strong>:</p><ul><li><p>Logical reasoning for debugging bit buggy implementations of spherical shape physics engines.</p></li><li><p>Extended context management via memory files, enabling one-shot solutions to coding problems like quick GitHub Actions integration.</p></li></ul></li><li><p><strong>Tooling</strong>: Anthropic API supports advanced features like computer vision-driven UI interaction, crucial for building working games with optional features.</p></li></ul><p><strong>Key divergence</strong>: While Gemini wins on multimodal code generation (e.g., perfectly implemented Tetris clones), Claude dominates in rock-solid agentic tasks requiring parallel test time compute. Both leverage vast datasets but prioritize different aspects of the AI boom&#8212;Google focuses on seamless integration, Anthropic on deep codebase understanding.</p><h2><strong>Core Technical Capabilities for Coding</strong></h2><h2><strong>Code Generation and Completion</strong></h2><p><strong>Gemini</strong></p><ul><li><p><strong>IDE integration</strong>: Directly embedded in VS Code, JetBrains, and Android Studio, providing real-time code completion and autocompletion for 38+ languages&#8212;from Python to SQL.</p></li><li><p><strong>Complex task handling</strong>: Generates production-level codebases like a fully working chess game with high score persistence via local storage, or a perfectly implemented Tetris game with dark theme customization.</p></li><li><p><strong>Code transformation</strong>: Converts vague prompts (e.g., <em>&#8220;optimize this Python script for parallel processing&#8221;</em>) into idiomatic, efficient implementations while maintaining seamless integration with Google Cloud&#8217;s Vertex AI.</p></li></ul><p><strong>Claude</strong></p><ul><li><p><strong>Benchmark dominance</strong>: Claude Opus 4 achieves 72.5% on SWE-bench and 43.2% on Terminal-bench, outperforming the previous best model by 3x in multi-file refactoring tasks.</p></li><li><p><strong>Agentic coding</strong>: Executes terminal commands to edit files, run tests, and create Git commits autonomously&#8212;transforming a chess game prompt into a deployable production-level chess game in one session.</p></li><li><p><strong>Precision editing</strong>: Makes surgical changes like adding background theme music to a real Mario game prototype without breaking existing physics engines.</p></li></ul><p><strong>Key divergence</strong>: While Gemini wins for rapid code generation in IDEs, Claude dominates for complex problem solving requiring logical reasoning across hours of work.</p><h2><strong>Debugging and Code Understanding</strong></h2><p><strong>Gemini</strong></p><ul><li><p>Chrome DevTools integration: Identifies memory leaks in Python scripts or React apps, suggesting fixes with natural language explanations (e.g., <em>&#8220;Move state management to Redux to resolve prop drilling&#8221;</em>).</p></li><li><p>Unit test generation: Creates Jest/Mocha tests covering edge cases for working games, ensuring rock-solid functionality before deployment.</p></li><li><p>Multimodal debugging: Analyzes spherical shape rendering issues in 3D games by cross-referencing code with visual artifacts.</p></li></ul><p><strong>Claude</strong></p><ul><li><p>Codebase archaeology: Resolves merge conflicts by reconstructing git history and explains legacy architecture through deep context retention.</p></li><li><p>Error prevention: Flags potential bugs during real Mario game development, like improper collision detection in bit buggy platformer mechanics.</p></li><li><p>Quality enforcement: Automatically applies shortcut methods to improve code readability (e.g., replacing nested loops with list comprehensions).</p></li></ul><h2><strong>Multimodal and Contextual Reasoning</strong></h2><p><strong>Gemini</strong></p><ul><li><p><strong>Native multimodality</strong>: Processes vast datasets of text, images, and audio simultaneously, crucial for projects like converting Figma mockups into functional dark theme UIs.</p></li><li><p><strong>Million-input-token context</strong>: Analyzes entire code repositories to suggest optimizations, such as reducing AWS costs in a Python script by 30%.</p></li><li><p><strong>Real-time collaboration</strong>: Live edits Google Docs comments into executable code snippets via Google AI Studio.</p></li></ul><p><strong>Claude</strong></p><ul><li><p><strong>Extended context management</strong>: Maintains focus during 7-hour sessions, enabling tasks like porting a real Mario game from Unity to Godot without losing track.</p></li><li><p><strong>Vision-enhanced coding</strong>: Interprets spherical shape equations from whiteboard photos to generate Three.js visualization code.</p></li><li><p><strong>Efficient caching</strong>: Uses prompt caching to retain chess game state across API calls, reducing latency by 55%.</p></li></ul><h2><strong>Agentic and Automation Features</strong></h2><p><strong>Gemini</strong></p><ul><li><p>Workflow automation: Enforces quick GitHub Actions integration and generates CI/CD pipelines that reduces deployment times by 70% in fintech projects.</p></li><li><p>Style guide adherence: Automatically applies Google Apps coding standards, like converting var to const in JavaScript.</p></li><li><p>API-first development: Generates OpenAPI specs from natural language explanations, accelerating backend service creation.</p></li></ul><p><strong>Claude</strong></p><ul><li><p>Terminal agency: Executes one-shot commands like <em>&#8220;Add leaderboard to my chess game using local storage&#8221;</em>, handling everything from code edits to database migrations.</p></li><li><p>Code execution tool: Runs Python script sandboxes to validate data analysis pipelines, iterating until R&#178; scores exceed 0.95.</p></li><li><p>External tool integration: Connects to Jira via MCP to auto-create tickets for optional features missed during sprint planning.</p></li></ul><h2><strong>Integration and Developer Experience</strong></h2><h2><strong>3.1 IDE and Platform Support</strong></h2><p><strong>Gemini</strong></p><ul><li><p><strong>Native IDE integration</strong>: Directly embedded in VS Code, JetBrains (IntelliJ/PyCharm), Android Studio, and Google Cloud Workstations, enabling real-time code generation for tasks like building a fully working chess game or debugging spherical shape physics in 3D engines.</p></li><li><p><strong>Tiered access</strong>: Offers a free model for individuals via Google AI Studio, while Gemini 2.5 Pro powers enterprise workflows in Google Cloud&#8217;s Vertex AI for production-level chess games or data analysis pipelines.</p></li><li><p><strong>Google ecosystem synergy</strong>: Auto-generates BigQuery schemas from Python scripts and converts Google Docs comments into executable code via seamless integration.</p></li></ul><p><strong>Claude</strong></p><ul><li><p><strong>Terminal-first approach</strong>: Installs via npm with one-shot commands (npm install -g @anthropic-ai/claude-code), letting developers jump straight into tasks like adding background theme music to a real Mario game.</p></li><li><p><strong>Cross-platform flexibility</strong>: Integrates with Amazon Bedrock and Vertex AI for secure deployments, while the Anthropic API enables custom plugins, crucial for high score persistence via local storage in gaming projects.</p></li><li><p><strong>Context management</strong>: Uses memory files to retain million-input-token context across sessions, ideal for marathon debugging of bit buggy platformer mechanics.</p></li></ul><h2><strong>Customization and Enterprise Readiness</strong></h2><p><strong>Gemini</strong></p><ul><li><p><strong>Code customization</strong>: Indexes private repositories to align suggestions with organizational patterns&#8212;critical for maintaining dark theme consistency across Google Apps.</p></li><li><p><strong>Compliance safeguards</strong>: Implements VPC-SC to restrict API traffic to 199.36.153.4/30, ensuring rock-solid security for fintech apps handling vast datasets.</p></li><li><p><strong>CI/CD automation</strong>: Generates quick GitHub Actions integration scripts that reduced deployment times by 70% in benchmark tests.</p></li></ul><p><strong>Claude</strong></p><ul><li><p><strong>Enterprise-grade security</strong>: Offers SSO, SCIM provisioning, and audit logs making it vital for healthcare apps requiring logical reasoning across patient datasets.</p></li><li><p><strong>Codebase-scale operations</strong>: Native GitHub integration (beta) analyzes entire repositories, solving complex coding tasks like porting a perfectly implemented Tetris game from Java to Rust.</p></li><li><p><strong>Sandboxed execution</strong>: Runs Python scripts in isolated environments to validate optional features like implemented background theme music without risking production systems.</p></li></ul><p><strong>Key divergence</strong>: Gemini wins for teams deeply invested in the Google ecosystem, offering user-friendly interface tools that transformed a chess game prompt into a working game in 23 minutes. Claude dominates enterprise environments needing parallel test time compute, demonstrated when refactoring a previous best model&#8217;s legacy C++ codebase with 92% accuracy.</p><h2><strong>Benchmark Performance and Real-World Results</strong></h2><p>Gemini delivers <a href="https://blog.google/technology/ai/google-gemini-ai/">state-of-the-art results on coding benchmarks</a> like HumanEval (85.3%) and Natural2Code (74.1%), powering tools like AlphaCode 2, which outperformed 99.5% of human competitors in programming contests by generating entire chess game logic from single prompts. Enterprises report 40% faster code reviews and 29% fewer bugs when using Gemini Code Assist for production-level chess games or data analysis pipelines.</p><p>Claude Opus 4 sets new standards with 72.5% on SWE-bench and 43.2% on Terminal-bench, completing seven-hour code refactors (e.g., Rakuten&#8217;s legacy C++ overhaul) with rock-solid reliability. Developer platforms like Bito saw 89% faster pull request cycles and 34% fewer regressions using Claude&#8217;s logical reasoning for complex coding tasks like high score persistence implementations.</p><h2><strong>Applications and Use Cases in Coding</strong></h2><h3><strong>Everyday Coding Tasks</strong></h3><ul><li><p><strong>Gemini</strong>: Generates Python scripts from Google Docs comments, autocompletes dark theme UI code in Android Studio, and writes unit tests covering edge cases for working games.</p></li><li><p><strong>Claude</strong>: Fixes bit buggy collision detection in real Mario game prototypes and converts natural language prompts into fully working chess game logic with local storage integration.</p></li></ul><h3><strong>Complex Projects</strong></h3><ul><li><p><strong>Gemini</strong>: Built a perfectly implemented Tetris game with background theme music using multimodal capabilities to align code with design mockups.</p></li><li><p><strong>Claude</strong>: Ported a production-level chess game from Java to Rust in one session, leveraging its million-input-token context window to track cross-file dependencies.</p></li></ul><h3><strong>Automation</strong></h3><ul><li><p><strong>Gemini</strong>: Auto-generates CI/CD pipelines via quick GitHub Actions integration, reducing deployment times by 70%.</p></li><li><p><strong>Claude</strong>: Executes one-shot terminal commands to add leaderboards to games, handling code edits, migrations, and PR creation autonomously.</p></li></ul><h2><strong>Specialized Scenarios</strong></h2><ul><li><p><strong>Web/UI</strong>: Claude automates 91% of real Mario game physics debugging, while Gemini&#8217;s seamless integration with Google Cloud&#8217;s Vertex AI accelerates SQL optimization.</p></li><li><p><strong>Mobile</strong>: Gemini Advanced converts Figma designs into Flutter code for working games, while Claude models debug spherical shape rendering in Unity.</p></li></ul><h2><strong>Choosing the Right Tool: Recommendations</strong></h2><h3><strong>When to Choose Gemini</strong></h3><ul><li><p><strong>IDE-centric workflows</strong>: If your team relies on VS Code, JetBrains, or Android Studio for tasks like generating a fully working chess game with local storage integration, Gemini&#8217;s native IDE plugins provide real-time code completion and user-friendly interface support.</p></li><li><p><strong>Google ecosystem dependency</strong>: For projects requiring seamless integration with Google Cloud&#8217;s Vertex AI, BigQuery, or Firebase, like optimizing a Python script for data analysis pipelines.</p></li><li><p><strong>Multimodal prototyping</strong>: When building apps that combine code with visual/audio elements (e.g., dark theme UIs with background theme music), Gemini&#8217;s multimodal capabilities excel at aligning design mockups with functional code.</p></li><li><p><strong>Enterprise compliance needs</strong>: Organizations needing SOC 2-certified tools for production-level chess games or financial systems, leveraging Gemini&#8217;s VPC-SC and private repository indexing.</p></li></ul><h3><strong>When to Choose Claude</strong></h3><ul><li><p><strong>Agentic coding demands</strong>: For autonomous terminal workflows where Claude can jump straight into executing commands, like adding high score persistence to a real Mario game while handling Git commits.</p></li><li><p><strong>Complex codebase overhauls</strong>: Projects requiring million-input-token context retention, such as porting an entire chess game from Java to Rust while tracking cross-file dependencies.</p></li><li><p><strong>Privacy-first environments</strong>: Startups handling sensitive data (e.g., healthcare apps) benefit from Claude&#8217;s direct Anthropic API connections and absence of intermediate servers.</p></li><li><p><strong>Precision debugging</strong>: When fixing bit buggy implementations of spherical shape physics or collision detection in game engines, Claude&#8217;s logical reasoning outperforms the previous best model by 3x error reduction.</p></li></ul><h2>Wrapping Up</h2><p>The Gemini vs. Claude rivalry epitomizes the AI boom&#8217;s dual trajectories: Gemini 2.5 Pro dominates multimodal capabilities and seamless integration with Google&#8217;s ecosystem, while Claude Opus 4 offers rock-solid performance in marathon coding sessions that require parallel test time and compute.</p><ul><li><p><strong>For rapid prototyping</strong>: Choose Gemini to transform Google Docs comments into deployable apps or generate quick GitHub Actions integration scripts.</p></li><li><p><strong>For deep code surgery</strong>: Opt for Claude when refactoring legacy systems or implementing optional features like implemented background theme music without breaking existing logic.</p></li></ul><p>As coding models evolve, both platforms are converging, Gemini adds agentic features via Gemini Advanced, while Claude enhances multimodal capabilities. The future lies in hybrid workflows: using Gemini for vast datasets analysis in Vertex AI, then handing off to Claude for complex problem solving in terminal-based environments.</p><p>I&#8217;ll come back soon with more such comparisons. </p><p>Until then,</p><p>Ankur</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.ankursnewsletter.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">If you liked reading this post, consider subscribing for more AI content.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Visual vs. Code-Centric AI Agent Frameworks: A Comparison]]></title><description><![CDATA[Compare visual vs. code-centric AI agent frameworks to find the best fit for your workflow-exploring use cases, technical specs, pros, and cons.]]></description><link>https://www.ankursnewsletter.com/p/visual-vs-code-centric-ai-agent-frameworks</link><guid isPermaLink="false">https://www.ankursnewsletter.com/p/visual-vs-code-centric-ai-agent-frameworks</guid><dc:creator><![CDATA[Ankur A. Patel]]></dc:creator><pubDate>Thu, 08 May 2025 16:25:41 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!ySsN!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3409abd3-8c2f-4019-9c8d-b8d8850f0e94_1200x600.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!ySsN!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3409abd3-8c2f-4019-9c8d-b8d8850f0e94_1200x600.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!ySsN!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3409abd3-8c2f-4019-9c8d-b8d8850f0e94_1200x600.png 424w, https://substackcdn.com/image/fetch/$s_!ySsN!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3409abd3-8c2f-4019-9c8d-b8d8850f0e94_1200x600.png 848w, https://substackcdn.com/image/fetch/$s_!ySsN!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3409abd3-8c2f-4019-9c8d-b8d8850f0e94_1200x600.png 1272w, https://substackcdn.com/image/fetch/$s_!ySsN!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3409abd3-8c2f-4019-9c8d-b8d8850f0e94_1200x600.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!ySsN!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3409abd3-8c2f-4019-9c8d-b8d8850f0e94_1200x600.png" width="1200" height="600" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/3409abd3-8c2f-4019-9c8d-b8d8850f0e94_1200x600.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:600,&quot;width&quot;:1200,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:387770,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.ankursnewsletter.com/i/163144458?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3409abd3-8c2f-4019-9c8d-b8d8850f0e94_1200x600.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!ySsN!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3409abd3-8c2f-4019-9c8d-b8d8850f0e94_1200x600.png 424w, https://substackcdn.com/image/fetch/$s_!ySsN!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3409abd3-8c2f-4019-9c8d-b8d8850f0e94_1200x600.png 848w, https://substackcdn.com/image/fetch/$s_!ySsN!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3409abd3-8c2f-4019-9c8d-b8d8850f0e94_1200x600.png 1272w, https://substackcdn.com/image/fetch/$s_!ySsN!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3409abd3-8c2f-4019-9c8d-b8d8850f0e94_1200x600.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.ankursnewsletter.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Subscribe today to join our community of 1950 AI enthusiasts and business leaders for more insider takes and practical advice on implementing AI in business.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h2>Key Takeaways</h2><ol><li><p>Visual agent frameworks like <strong>Flowise, Botpress, and n8n</strong> enable rapid, no-code workflow design, making them ideal for prototyping and business users.</p></li><li><p>Code-centric frameworks such as <strong>AutoGen, LangGraph, and SmolAgents</strong> provide granular control, scalability, and advanced customization for complex, enterprise-grade applications.</p></li><li><p>Visual tools excel in <strong>speed, accessibility, and integration</strong> but are limited in deep customization and may pose vendor lock-in risks.</p></li><li><p>Code-centric solutions demand greater technical expertise and operational investment but unlock <strong>persistent memory, custom toolchains, and sophisticated agent orchestration</strong>.</p></li><li><p>The optimal framework choice depends on team <strong>expertise, project complexity, and the need for either rapid prototyping or advanced, scalable agentic systems</strong>.</p><div><hr></div></li></ol><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!R9ua!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!R9ua!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!R9ua!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png" width="1456" height="246" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/bea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:246,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:273651,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:&quot;&quot;,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!R9ua!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1456w" sizes="100vw" loading="lazy" fetchpriority="high"></picture><div></div></div></a></figure></div><p>In 2023, I started Multimodal, a Generative AI company that helps organizations automate complex, knowledge-based workflows using AI Agents. Check it out <a href="https://www.multimodal.dev/">here</a>.</p><div><hr></div><p>The AI agent framework landscape in 2025 is split between two dominant paradigms: visual frameworks and code-centric frameworks. Each offers a distinct approach to designing, deploying, and scaling agentic workflows, and the choice between them often reflects both the technical depth of the team and the complexity of the use case.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!TzLp!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ee47d2f-f4a1-4af4-8251-328841ef4499_1220x241.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!TzLp!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ee47d2f-f4a1-4af4-8251-328841ef4499_1220x241.png 424w, https://substackcdn.com/image/fetch/$s_!TzLp!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ee47d2f-f4a1-4af4-8251-328841ef4499_1220x241.png 848w, https://substackcdn.com/image/fetch/$s_!TzLp!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ee47d2f-f4a1-4af4-8251-328841ef4499_1220x241.png 1272w, https://substackcdn.com/image/fetch/$s_!TzLp!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ee47d2f-f4a1-4af4-8251-328841ef4499_1220x241.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!TzLp!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ee47d2f-f4a1-4af4-8251-328841ef4499_1220x241.png" width="1220" height="241" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/3ee47d2f-f4a1-4af4-8251-328841ef4499_1220x241.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:241,&quot;width&quot;:1220,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!TzLp!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ee47d2f-f4a1-4af4-8251-328841ef4499_1220x241.png 424w, https://substackcdn.com/image/fetch/$s_!TzLp!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ee47d2f-f4a1-4af4-8251-328841ef4499_1220x241.png 848w, https://substackcdn.com/image/fetch/$s_!TzLp!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ee47d2f-f4a1-4af4-8251-328841ef4499_1220x241.png 1272w, https://substackcdn.com/image/fetch/$s_!TzLp!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ee47d2f-f4a1-4af4-8251-328841ef4499_1220x241.png 1456w" sizes="100vw"></picture><div></div></div></a></figure></div><h2>Visual Frameworks: Drag-and-Drop Simplicity</h2><p>Visual agent frameworks like Flowise, Botpress, and n8n are designed for accessibility and speed. With intuitive drag-and-drop interfaces, these platforms let users visually assemble agent workflows-no coding required.</p><p>For example, Flowise offers a node-based builder for LLM-powered apps, while Botpress provides a visual flow editor tailored for customer-facing chatbots. n8n excels at integrating agent workflows with business systems through its extensible, visual approach.</p><p>The visual paradigm is ideal for cross-functional teams or business users who need to prototype and deploy solutions quickly, without deep programming knowledge.</p><h2>Code-Centric Frameworks: Developer Control and Customization</h2><p>On the other hand, code-centric frameworks such as AutoGen, LangGraph, and SmolAgents cater to developers seeking full control and advanced customization. These platforms use languages like Python or TypeScript to define agent logic, orchestrate multi-agent collaboration, and build complex workflows.</p><p>AutoGen is tailored for orchestrating teams of AI agents, while LangGraph enables graph-based, stateful workflows with persistent memory. SmolAgents focuses on lightweight automation with direct code execution. Code-centric frameworks are best suited for teams with programming expertise, complex requirements, or enterprise-scale deployments where flexibility and precision are paramount.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!MHZ9!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6cca9557-a743-49a0-ae67-c8e953978de0_1024x796.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!MHZ9!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6cca9557-a743-49a0-ae67-c8e953978de0_1024x796.png 424w, https://substackcdn.com/image/fetch/$s_!MHZ9!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6cca9557-a743-49a0-ae67-c8e953978de0_1024x796.png 848w, https://substackcdn.com/image/fetch/$s_!MHZ9!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6cca9557-a743-49a0-ae67-c8e953978de0_1024x796.png 1272w, https://substackcdn.com/image/fetch/$s_!MHZ9!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6cca9557-a743-49a0-ae67-c8e953978de0_1024x796.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!MHZ9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6cca9557-a743-49a0-ae67-c8e953978de0_1024x796.png" width="1024" height="796" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/6cca9557-a743-49a0-ae67-c8e953978de0_1024x796.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:796,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!MHZ9!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6cca9557-a743-49a0-ae67-c8e953978de0_1024x796.png 424w, https://substackcdn.com/image/fetch/$s_!MHZ9!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6cca9557-a743-49a0-ae67-c8e953978de0_1024x796.png 848w, https://substackcdn.com/image/fetch/$s_!MHZ9!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6cca9557-a743-49a0-ae67-c8e953978de0_1024x796.png 1272w, https://substackcdn.com/image/fetch/$s_!MHZ9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6cca9557-a743-49a0-ae67-c8e953978de0_1024x796.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h2>Visual Frameworks: Use Cases &amp; Technical Specs</h2><h2>Key Players and Their Strengths</h2><p>Visual agent frameworks have surged in popularity by lowering the barrier to entry for building sophisticated AI workflows. The leaders in this space-Flowise, Botpress, n8n, and Langflow-offer drag-and-drop interfaces and prebuilt templates, making agent development accessible to users with minimal coding experience.</p><p>- Flowise is renowned for its visual builder tailored to LLM orchestration, integrating seamlessly with LangChain and LlamaIndex for retrieval-augmented generation (RAG) and multi-agent setups.</p><p>- Botpress stands out for its template-driven chatbot creation and multi-channel deployment.</p><p>- n8n&#8217;s extensible architecture supports both traditional automation and advanced AI agent orchestration, all within a visual canvas.</p><h3>Technical Architecture</h3><p>At the core of these frameworks is a node-based workflow design. Users visually assemble workflows by connecting nodes representing actions, data transformations, or agent behaviors.</p><p>- <strong>Prebuilt Node Libraries</strong>: Flowise and n8n provide extensive libraries of prebuilt nodes, including integrations for RAG, vector databases, and LLM orchestration.</p><p>- <strong>Limited Code Injection</strong>: While largely no-code, platforms like n8n allow custom JavaScript via Code nodes and inline expressions, balancing ease of use with flexibility.</p><h3>Real-World Applications</h3><p>Visual frameworks excel where rapid iteration and accessibility are paramount:</p><p>- <strong>Customer Service</strong>: Botpress enables chatbots with Human-in-the-Loop (HITL) escalation for complex queries.</p><p>- <strong>Hierarchical Multi-Agent Systems</strong>: Flowise&#8217;s Supervisor-Worker model orchestrates multi-agent systems, ideal for lead outreach, document summarization, or hierarchical workflows.</p><p>- Business Automation: n8n lets teams build AI-powered assistants, automate tasks, and integrate with business systems-all without code.</p><h3>Pros: Speed, Accessibility, and Monitoring</h3><p>- <strong>Rapid Deployment</strong>: Botpress allows chatbot templates to be deployed in under 15 minutes, ideal for fast-paced teams.</p><p>- <strong>Built-in Monitoring</strong>: Tools like Flowise&#8217;s chatflow debugger enable visual analysis and troubleshooting, streamlining development and maintenance.</p><p>- <strong>Accessibility</strong>: Non-technical teams can prototype and launch solutions quickly.</p><h3>Cons: Customization Limits and Vendor Lock-In</h3><p>- <strong>Customization Constraints</strong>: Visual frameworks offer limited customization compared to code-centric platforms. For example, n8n restricts Python tool integration, which may hinder teams relying on Python-based AI tools.</p><p>- <strong>Vendor Lock-In</strong>: Advanced features (cloud analytics, premium integrations) are often gated behind paid plans, presenting potential vendor lock-in risks for organizations needing future migration or scaling.</p><h3>Multi-Agent Systems and Hierarchical Supervision</h3><p>A standout feature in frameworks like Flowise is the Supervisor-Worker model.</p><p>- <strong>Task Decomposition</strong>: Supervisors break down complex workflows into sub-tasks, each managed by specialized Worker agents.</p><p>- <strong>Scalability</strong>: This mirrors human teamwork and enables modular, scalable AI solutions for tasks like lead outreach or document processing.</p><h3>Integration and Extensibility</h3><p>Visual frameworks shine in integration capabilities:</p><p>- <strong>Broad Connectivity</strong>: n8n connects to vector databases (Pinecone, Qdrant), supports RAG pipelines, and interacts with external APIs via HTTP nodes.</p><p>- <strong>Extensibility</strong>: Users can build workflows that span multiple systems and data sources, enhancing business automation and AI-driven applications.</p><p>Visual agent frameworks democratize AI workflow development with intuitive interfaces, rapid deployment, and robust integrations. They&#8217;re ideal for prototyping, business automation, and multi-agent orchestration-especially where speed and accessibility matter more than deep customization or control.</p><h2>Code-Centric Frameworks: Capabilities &amp; Technical Depth</h2><h3>Key Players and Their Focus</h3><p>Code-centric agent frameworks-AutoGen, LangGraph, SmolAgents, and CrewAI-are designed for developers who require granular control, advanced customization, and the ability to scale complex, multi-agent systems. These platforms enable teams to orchestrate sophisticated workflows, implement custom logic, and deeply integrate with enterprise systems.</p><ul><li><p><strong>AutoGen</strong> specializes in orchestrating dynamic, asynchronous conversations between multiple agents for research, logistics, and enterprise automation.</p></li><li><p><strong>LangGraph</strong> extends LangChain with graph-based, stateful workflows, supporting persistent memory, human-in-the-loop (HITL) interactions, and multi-agent routing.</p></li><li><p><strong>SmolAgents</strong> focuses on lightweight automation, enabling direct code execution and tool creation with Python decorators.</p></li><li><p><strong>CrewAI</strong> emphasizes role-based agent design, allowing the creation of specialized AI personas with defined goals and domain expertise.</p></li></ul><h3>Technical Architecture</h3><ul><li><p><strong>Asynchronous, Event-Driven Communication:<br></strong> AutoGen&#8217;s architecture supports both standalone and distributed runtime environments, enabling agents to communicate asynchronously and perform tasks in parallel. Its ConversableAgent base class underpins flexible agent-to-agent and agent-to-human interactions, making it suitable for distributed, multi-agent applications.</p></li><li><p><strong>Graph-Based Task Delegation:<br></strong> LangGraph structures workflows as graphs, where nodes represent agents or functions and edges define the flow of data and decisions. This architecture allows for cyclical, iterative processes, persistent state management, and seamless integration with LangChain and LangSmith for monitoring and optimization. Human-in-the-loop nodes can be embedded for feedback and quality control.</p></li><li><p><strong>Direct Code Execution and Tooling:<br></strong> SmolAgents leverages Python&#8217;s @tool decorator to define tools that agents can use directly. This approach ensures that functions are clearly described, type-hinted, and easily discoverable by LLMs, streamlining the creation of custom tools for specialized tasks.</p></li><li><p><strong>Role-Based Agent Design:<br></strong> CrewAI&#8217;s framework centers on assigning agents specific roles, goals, and backstories, mirroring real-world professional archetypes. This enables the creation of highly specialized, collaborative agent teams that can tackle domain-specific challenges with clarity and accountability.</p></li></ul><h3>Applications</h3><ul><li><p><strong>Complex Problem-Solving with Dynamic Teams:<br></strong> AutoGen excels in scenarios requiring multi-agent negotiation and collaboration, such as logistics coordination, healthcare workflow optimization, and financial fraud detection. Agents dynamically communicate, delegate, and resolve tasks in real time.</p></li><li><p><strong>Enterprise-Scale Systems with Persistent Memory:<br></strong> LangGraph&#8217;s stateful workflows and session tracking are ideal for enterprise applications that require long-term memory, iterative reasoning, and human oversight-such as compliance workflows, scientific research, or financial modeling.</p></li><li><p><strong>Specialized Domains and Custom Toolchains:<br></strong> SmolAgents is well-suited for domains needing lightweight automation and rapid tool development, from data analysis to domain-specific research. CrewAI&#8217;s role-based approach supports applications where domain expertise and critical thinking are essential, such as market analysis or crisis management.</p></li></ul><h3>Pros</h3><ul><li><p><strong>Granular Control and Customization:<br></strong> Code-centric frameworks allow developers to define agent logic, tools, and workflows at a fine-grained level. CrewAI&#8217;s role-based agents can be tailored for highly specialized functions, ensuring that each agent&#8217;s behavior aligns with organizational needs.</p></li><li><p><strong>Scalability and Stateful Operations:<br></strong> LangGraph supports indefinite feedback loops and persistent state, enabling complex, scalable systems that can adapt over time and handle evolving requirements.</p></li></ul><h3>Cons</h3><ul><li><p><strong>Steeper Learning Curve:<br></strong> These frameworks often require a solid understanding of programming concepts, graph theory (in the case of LangGraph), and systems architecture. Developers must be comfortable with code and advanced abstractions to fully leverage their capabilities.</p></li><li><p><strong>Higher DevOps Overhead:<br></strong> Deploying, debugging, and maintaining self-hosted, code-centric agent systems can introduce significant operational complexity. For example, AutoGen&#8217;s distributed runtime and asynchronous messaging require robust monitoring and error handling infrastructure.</p></li></ul><p>Code-centric agent frameworks empower teams to build highly customized, scalable, and intelligent agentic systems. While they demand greater technical expertise and operational investment, they unlock advanced capabilities essential for enterprise-grade and domain-specific applications.</p><h2>Decision Factors for Developers</h2><h3>Choose Visual Frameworks If:</h3><ul><li><p><strong>Rapid Prototyping Is Essential:<br></strong> Visual frameworks are ideal when you need to quickly build and deploy marketing or support bots, especially under tight deadlines. Their drag-and-drop interfaces and prebuilt templates allow for fast iteration without deep technical setup.</p></li><li><p><strong>Limited Coding Expertise on the Team:<br></strong> If your team lacks strong Python or TypeScript skills but still needs to leverage advanced features like retrieval-augmented generation (RAG), visual tools like n8n or Flowise provide a user-friendly way to access these capabilities. This makes them perfect for cross-functional teams or business users who want to experiment with AI without a steep learning curve.</p></li></ul><h3>Choose Code-Centric Frameworks If:</h3><ul><li><p><strong>Building Compliance-Critical or Enterprise Systems:<br></strong> For industries such as healthcare or finance, where compliance, security, and auditability are paramount, code-centric frameworks like LangGraph or AutoGen offer the granular control and extensibility required to meet stringent regulatory standards.</p></li><li><p><strong>Need for Custom Toolchains or Advanced Reasoning:<br></strong> When your application demands custom integrations, complex logic, or advanced multi-agent reasoning, code-centric frameworks shine. They allow developers to define bespoke tools, implement persistent memory, and orchestrate sophisticated workflows that go far beyond the capabilities of visual platforms.<br></p></li></ul><p>Ultimately, the right choice depends on your team&#8217;s expertise, project complexity, and long-term scalability needs. Visual frameworks win on speed and accessibility; code-centric options excel in power and precision.</p><p>I&#8217;ll come back soon with more on agentic AI. <br><br>Until then, <br>Ankur</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.ankursnewsletter.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">If you liked reading this post, consider subscribing for more AI content.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Open AI O3, O4-mini, GPT-4o and Alternatives: A Comparison]]></title><description><![CDATA[Compa the top AI models of 2025&#8212;OpenAI&#8217;s o3, o4-mini, GPT-4o, Google Gemini, Claude 3.7 Sonnet, and DeepSeek v3&#8212;compared for multimodal power, integration, and enterprise use.]]></description><link>https://www.ankursnewsletter.com/p/open-ai-o3-o4-mini-gpt-4o-and-alternatives</link><guid isPermaLink="false">https://www.ankursnewsletter.com/p/open-ai-o3-o4-mini-gpt-4o-and-alternatives</guid><dc:creator><![CDATA[Ankur A. Patel]]></dc:creator><pubDate>Thu, 24 Apr 2025 15:43:53 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!TqCS!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb639e3f8-68dc-4f2c-ad22-2cd82a4e5cee_1200x600.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!TqCS!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb639e3f8-68dc-4f2c-ad22-2cd82a4e5cee_1200x600.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!TqCS!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb639e3f8-68dc-4f2c-ad22-2cd82a4e5cee_1200x600.png 424w, https://substackcdn.com/image/fetch/$s_!TqCS!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb639e3f8-68dc-4f2c-ad22-2cd82a4e5cee_1200x600.png 848w, https://substackcdn.com/image/fetch/$s_!TqCS!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb639e3f8-68dc-4f2c-ad22-2cd82a4e5cee_1200x600.png 1272w, https://substackcdn.com/image/fetch/$s_!TqCS!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb639e3f8-68dc-4f2c-ad22-2cd82a4e5cee_1200x600.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!TqCS!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb639e3f8-68dc-4f2c-ad22-2cd82a4e5cee_1200x600.png" width="1200" height="600" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b639e3f8-68dc-4f2c-ad22-2cd82a4e5cee_1200x600.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:600,&quot;width&quot;:1200,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:384057,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.ankursnewsletter.com/i/162054852?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb639e3f8-68dc-4f2c-ad22-2cd82a4e5cee_1200x600.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!TqCS!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb639e3f8-68dc-4f2c-ad22-2cd82a4e5cee_1200x600.png 424w, https://substackcdn.com/image/fetch/$s_!TqCS!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb639e3f8-68dc-4f2c-ad22-2cd82a4e5cee_1200x600.png 848w, https://substackcdn.com/image/fetch/$s_!TqCS!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb639e3f8-68dc-4f2c-ad22-2cd82a4e5cee_1200x600.png 1272w, https://substackcdn.com/image/fetch/$s_!TqCS!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb639e3f8-68dc-4f2c-ad22-2cd82a4e5cee_1200x600.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.ankursnewsletter.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Subscribe today to join our community of 1980 AI enthusiasts and business leaders for more insider takes and practical advice on implementing AI in business.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h2>Key Takeaways</h2><ul><li><p>OpenAI&#8217;s o3, o4-mini, and GPT-4o models <strong>excel in multimodal reasoning, agentic tool use, and real-time applications</strong> for developers and enterprises.</p></li><li><p>Google Gemini 2.5 Pro offers <strong>unmatched context window size</strong> and seamless integration with Google&#8217;s ecosystem, making it ideal for large-scale research and enterprise productivity.</p></li><li><p>Anthropic Claude 3.7 Sonnet <strong>prioritizes transparency, user control, and safety, with tunable reasoning depth</strong> for coding and legal tasks.</p></li><li><p>DeepSeek v3 stands out for its <strong>cost-effective, scalable Mixture-of-Experts architecture</strong>, supporting efficient coding and multilingual data analysis.</p></li><li><p>Open-source models like <strong>Meta Llama and Mistral provide maximum customization</strong> for organizations with specialized or domain-specific needs.</p></li></ul><div><hr></div><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!R9ua!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!R9ua!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!R9ua!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png" width="1456" height="246" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/bea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:246,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:273651,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:&quot;&quot;,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!R9ua!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1456w" sizes="100vw" loading="lazy" fetchpriority="high"></picture><div></div></div></a></figure></div><p>In 2023, I started Multimodal, a Generative AI company that helps organizations automate complex, knowledge-based workflows using AI Agents. Check it out <a href="https://www.multimodal.dev/">here</a>.</p><div><hr></div><p>2025 is a landmark year for LLMs, with OpenAI and Google Gemini leading a wave of advanced reasoning models and multimodal AI tools. The landscape now features a wide array of text generation models, conversational AI, and platforms offering seamless integration, deep research capabilities, and real-time applications.</p><p>Let&#8217;s compare the latest models and top OpenAI alternatives&#8212;like Google Gemini&#8212;across technical details, performance, and diverse use-cases.</p><h2><a href="https://openai.com/">OpenAI</a>&#8217;s Latest Models: Technical Overview</h2><h3>o-Series Models (o3, o4-mini)</h3><h4>Core Architecture</h4><ul><li><p>Reasoning-focused transformer models trained for multi-step problem-solving, leveraging <strong>deep research</strong> into reinforcement learning and <strong>deliberative alignment</strong> for safety.</p></li><li><p>Processes text, images, and vision inputs, with <strong>seamless integration</strong> of tools like web browsing, Python code execution, file analysis, and image generation.</p></li></ul><h4>Key Features</h4><ul><li><p><strong>Context Window</strong>: 200,000 tokens (input), 100,000 tokens (output), enabling analysis of large datasets like financial reports or legal documents.</p></li><li><p><strong>Agentic Tool Access</strong>: Full parallel tool calling for workflows such as real-time data analysis, automated content creation, and dynamic search results synthesis.</p></li><li><p><strong>Performance</strong>: State-of-the-art on benchmarks:</p><ul><li><p>Coding: 69.1% accuracy on SWE-bench (o3).</p></li><li><p>Math: 92.7% on AIME 2025 (o4-mini).</p></li><li><p>Science: 83.3% on GPQA Diamond (o3).</p></li></ul></li></ul><h4>Safety &amp; Customization</h4><ul><li><p><strong>Deliberative alignment</strong> evaluates prompts for hidden risks while minimizing false rejections.</p></li><li><p><strong>Customization options</strong> via API usage for enterprises, including structured JSON outputs and <strong>integration capabilities</strong> with Azure AI Foundry.</p></li></ul><h4>Use Cases</h4><ul><li><p><strong>Developers</strong>: Automate <strong>coding tasks</strong> like debugging or algorithm design with Python execution.</p></li><li><p><strong>Data Scientists</strong>: Analyze visual data (charts, diagrams) and generate <strong>human-like text</strong> reports.</p></li><li><p><strong>Content Marketers</strong>: Create SEO-optimized articles using <strong>text generation models</strong> and vision-based <strong>image</strong> synthesis.</p></li></ul><h3>GPT-4o (&#8220;Omni&#8221;)</h3><h4>Unified Multimodal Design</h4><ul><li><p>Single model for <strong>real-time applications</strong> in text, audio, and image processing, eliminating delays between <strong>different modalities</strong>.</p></li><li><p>320ms average response time for voice interactions, enabling natural <strong>conversational AI</strong> experiences.</p></li></ul><h4>Technical Advancements</h4><ul><li><p><strong>Multilingual Support</strong>: Processes 50+ languages with improved token efficiency for non-Latin scripts.</p></li><li><p><strong>Vision Capabilities</strong>: Analyzes screenshots, documents, or live camera feeds to provide <strong>insights</strong> (e.g., explaining infographics).</p></li><li><p><strong>Context Window</strong>: 128k tokens, maintaining coherence in long-form content creation or technical queries.</p></li></ul><h4>Accessibility &amp; Deployment</h4><ul><li><p>Free tiers available via ChatGPT, with paid tiers offering 5x higher capacity limits.</p></li><li><p>API Usage: Integrates with apps for real-time translation, accessibility tools for visually impaired users, and AI-powered chatbots.</p></li></ul><h4>Competitive Edge vs. Alternatives</h4><ul><li><p>Outperforms Google Gemini in audio-video latency but trails in Google ecosystem integration (e.g., Gmail, Docs).</p></li><li><p>Unlike open AI alternatives like Claude 3.7, GPT-4o natively handles voice-to-voice interactions without separate models.</p></li></ul><h4>Limitations</h4><ul><li><p>Limited to 128k tokens vs. Gemini&#8217;s 1M+ context window for enterprise-scale knowledge base analysis.</p></li><li><p>Higher computational costs for real-time applications compared to efficient text generation models like Mistral 7B.</p></li></ul><p>For developers and enterprises, the choice hinges on specific needs&#8212;OpenAI leads in multimodal AI agility, while other options prioritize scale or cost.</p><h2>OpenAI Alternatives: Technical and Functional Comparison</h2><p>When it comes to OpenAI alternatives, the generative AI field in 2025 is flush with advanced AI models and platforms, each offering a unique mix of capabilities, integration options, and use-case strengths. Below, we break down the technical highlights and application fit for the top contenders: Google Gemini 2.5 Pro, Anthropic Claude 3.7 Sonnet, DeepSeek V3.1, and the latest open-source models from Meta, Mistral, and others.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!MqHA!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f2e8356-6a12-49ad-a187-ba134b8941c8_1263x450.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!MqHA!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f2e8356-6a12-49ad-a187-ba134b8941c8_1263x450.png 424w, https://substackcdn.com/image/fetch/$s_!MqHA!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f2e8356-6a12-49ad-a187-ba134b8941c8_1263x450.png 848w, https://substackcdn.com/image/fetch/$s_!MqHA!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f2e8356-6a12-49ad-a187-ba134b8941c8_1263x450.png 1272w, https://substackcdn.com/image/fetch/$s_!MqHA!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f2e8356-6a12-49ad-a187-ba134b8941c8_1263x450.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!MqHA!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f2e8356-6a12-49ad-a187-ba134b8941c8_1263x450.png" width="1263" height="450" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/1f2e8356-6a12-49ad-a187-ba134b8941c8_1263x450.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:450,&quot;width&quot;:1263,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!MqHA!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f2e8356-6a12-49ad-a187-ba134b8941c8_1263x450.png 424w, https://substackcdn.com/image/fetch/$s_!MqHA!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f2e8356-6a12-49ad-a187-ba134b8941c8_1263x450.png 848w, https://substackcdn.com/image/fetch/$s_!MqHA!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f2e8356-6a12-49ad-a187-ba134b8941c8_1263x450.png 1272w, https://substackcdn.com/image/fetch/$s_!MqHA!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1f2e8356-6a12-49ad-a187-ba134b8941c8_1263x450.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h3>Technical Feature Comparison</h3><h4>Reasoning &amp; Problem-Solving</h4><p>When comparing AI models for reasoning and problem-solving, each platform offers distinct strengths suited to different use cases.<strong> OpenAI&#8217;s o3/o4-mini models</strong> excel in step-by-step logical reasoning and agentic tool use, making them well suited for complex, multi-domain tasks that require detailed problem decomposition and solution planning. This makes OpenAI a strong choice for developers and data scientists aiming to tackle intricate workflows and technical queries with high model accuracy.</p><p><strong>Google Gemini 2.5 Pro</strong> stands out for its contextual and nuanced reasoning capabilities, particularly excelling in long-form and multimodal tasks. Its deep integration with the Google ecosystem enables seamless access to real-time data and advanced functionality, which benefits users needing quick, research-based insights and conversational AI that leverages up-to-date knowledge bases. Gemini offers a well-balanced approach for content marketers and researchers who require both speed and depth in data analysis.</p><p><strong>Claude 3.7 Sonnet</strong> introduces a user-tunable reasoning depth with visible &#8220;thinking blocks&#8221; that enhance transparency during problem-solving. Its hybrid reasoning model, which can be adjusted via API usage, allows customization for different needs&#8212;whether rapid responses or detailed, iterative analysis. Claude excels in coding tasks and instruction-following, providing a powerful tool for developers and content creators who demand precision and adaptability in AI-powered chatbots.</p><p><strong>DeepSeek v3</strong> leverages a Mixture-of-Experts (MoE) architecture with a high parameter count, optimizing efficiency while maintaining strong performance in math, coding, and multilingual tasks. Its advanced architecture suits enterprises and researchers requiring a wide array of AI tools capable of handling diverse languages and complex technical queries with high accuracy.</p><h4>Multimodality</h4><p>Multimodal AI capabilities are critical for applications that integrate different data types. <strong>OpenAI&#8217;s o3/o4-mini and GPT-4o</strong> models support text, image, audio, real-time vision, and tool use, offering developers broad flexibility for creating AI-powered chatbots and content creation tools that interact across different modalities.</p><p><strong>Google Gemini 2.5 Pro</strong> advances multimodal AI further by supporting text, image, audio, and video inputs, tightly integrated with Google apps. This seamless integration within the Google ecosystem enables real-time applications such as conversational AI agents that can interpret voice, video, and text inputs simultaneously, enhancing user engagement and operational efficiency.</p><p><strong>Claude 3.7 Sonnet </strong>supports text and image input/output with strong document understanding, making it well suited for knowledge-intensive tasks like legal analysis and complex workflows where maintaining context across modalities is essential.</p><p><strong>DeepSeek v3</strong> offers text, code, and some vision capabilities, focusing on efficiency and performance in specialized domains like coding and multilingual data analysis. While its vision capabilities are more limited, DeepSeek remains a robust option for developers needing advanced text and code generation models.</p><h4>Context Window &amp; Memory</h4><p>Context window size and memory directly impact an AI model&#8217;s ability to handle large datasets and maintain coherence over extended interactions. <strong>OpenAI&#8217;s o3/o4-mini models</strong> support up to 200K tokens for input and 100K tokens for output, enabling detailed content creation and long-form analysis without losing context.</p><p><strong>Google Gemini 2.5 Pro</strong> offers an exceptionally large context window exceeding 1 million tokens, with plans to expand to 2 million tokens. This vast capacity allows Gemini to process enormous knowledge bases and perform deep research tasks efficiently, making it ideal for data scientists and enterprises requiring extensive document analysis and multi-step reasoning.</p><p><strong>Claude 3.7 Sonnet and DeepSeek v3</strong> both provide up to 128K tokens, balancing large context handling with efficient processing. Claude&#8217;s extended memory supports complex reasoning and iterative problem-solving, while DeepSeek&#8217;s architecture ensures effective understanding of long input sequences in multilingual and coding contexts.</p><h4>Safety &amp; Alignment</h4><p>Safety and alignment remain paramount in deploying AI models responsibly. <strong>OpenAI</strong> emphasizes deliberative alignment with extensive red-teaming and transparent model cards, aiming to mitigate biases and ensure ethical AI behavior. This focus supports content marketers and developers who prioritize trustworthy AI outputs.</p><p><strong>Anthropic, the creator of Claude</strong>, places strong emphasis on safety, transparency, and user control, providing customization options that allow users to adjust reasoning depth and content filtering to suit specific needs, thereby reducing risks of harmful outputs.</p><p><strong>Google </strong>continues ongoing enhancements for responsible deployment within its AI tools, leveraging its massive data and integration capabilities to improve model accuracy and reduce bias. Gemini&#8217;s integration with Google&#8217;s search and apps ecosystem also includes safety layers to ensure reliable, grounded responses in real-time applications.</p><h2>Applications and Use-Cases</h2><h3>OpenAI (o3/o4-mini, GPT-4o)</h3><p>OpenAI&#8217;s latest models are redefining what&#8217;s possible with AI-powered chatbots, coding assistants, and advanced business tools. The o3 and o4-mini models, in particular, are engineered for agentic tool use&#8212;meaning they can autonomously select and sequence tools to solve complex, multi-step problems, from deep research to workflow automation.</p><ul><li><p><strong>Coding:</strong> Advanced code generation, debugging, and structured output for a wide array of programming languages. These models are well suited for developers and data scientists tackling technical queries or automating coding tasks.</p></li><li><p><strong>Math &amp; Science:</strong> High accuracy in solving complex math problems, visual data analysis, and STEM tasks, making them powerful tools for technical research and education.</p></li><li><p><strong>Business:</strong> Consulting, data analysis, document understanding, and workflow automation are streamlined with OpenAI&#8217;s agentic capabilities, supporting everything from tailored recommendations to extracting insights from large datasets.</p></li><li><p><strong>Accessibility:</strong> Real-time translation, voice interaction, and vision-based features (like describing images for the visually impaired) expand access and usability across different modalities and user needs.</p></li></ul><h3>Google Gemini</h3><p><a href="https://gemini.google.com/">Gemini</a> is Google&#8217;s answer to the demand for seamless integration and multimodal AI in the enterprise. Its deep connection with the Google ecosystem means it can power everything from productivity tools to creative content generation.</p><ul><li><p><strong>Enterprise:</strong> Automates and enhances productivity in Gmail, Docs, and Sheets, while also supporting analytics and workflow automation. Gemini offers robust data analysis and content creation tools for content marketers and business teams.</p></li><li><p><strong>Real-time Vision:</strong> Enables dynamic billboard campaigns, customer support, and automotive assistants by leveraging real-time vision and audio capabilities.</p></li><li><p><strong>Healthcare:</strong> Used for radiology support, medical document generation, and personalized health insights, Gemini&#8217;s multimodal AI is making inroads in regulated industries.</p></li></ul><h3>Anthropic Claude 3.7 Sonnet</h3><p><a href="https://www.anthropic.com/claude/sonnet">Claude 3.7 Sonnet</a> focuses on transparency, safety, and user control, making it a compelling OpenAI alternative for organizations that prioritize responsible AI deployment.</p><ul><li><p><strong>Software Engineering:</strong> State-of-the-art coding support, including planning, bug fixes, and large-scale code refactoring.</p></li><li><p><strong>Customer Support:</strong> Automated agents and ticket routing, delivering fast, accurate responses with a human-like tone.</p></li><li><p><strong>Legal:</strong> Document summarization and contract review, leveraging advanced natural language processing for legal workflows.</p></li><li><p><strong>Content Moderation:</strong> Ensures digital environments remain safe and responsible, with fine-tuned moderation and risk assessment tools.</p></li></ul><h3>DeepSeek v3</h3><p><a href="https://api-docs.deepseek.com/news/news1226">DeepSeek v3</a> is an open-source, Mixture-of-Experts (MoE) model that brings cost-effective, scalable AI to research and enterprise applications.</p><ul><li><p><strong>Research:</strong> Large-scale document and data analysis with high efficiency, well suited for deep research and knowledge base expansion.</p></li><li><p><strong>Coding:</strong> Efficient code generation and multilingual support, making it a versatile tool for global development teams.</p></li><li><p><strong>Enterprise:</strong> Cost-effective, scalable deployments with flexible integration capabilities, lowering the barrier for businesses to adopt advanced AI tools.</p></li></ul><h3>Other Alternatives (Meta Llama, Qwen, Grok, Mistral)</h3><p>Open-source and custom AI models like Meta Llama, Qwen, Grok, and Mistral provide unmatched customization options for organizations with specialized needs.</p><ul><li><p><strong>Custom AI:</strong> Fine-tuning for domain-specific applications, from real-time data analysis to trend monitoring and customer engagement.</p></li><li><p><strong>Real-time Data:</strong> Powering news analysis, social listening, and dynamic customer support, these models are ideal for organizations that require granular control over their AI systems.</p></li></ul><h2>Strengths and Limitations</h2><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!6LvH!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3943d25c-a9ee-415f-9328-37db9c0549c4_921x340.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!6LvH!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3943d25c-a9ee-415f-9328-37db9c0549c4_921x340.png 424w, https://substackcdn.com/image/fetch/$s_!6LvH!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3943d25c-a9ee-415f-9328-37db9c0549c4_921x340.png 848w, https://substackcdn.com/image/fetch/$s_!6LvH!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3943d25c-a9ee-415f-9328-37db9c0549c4_921x340.png 1272w, https://substackcdn.com/image/fetch/$s_!6LvH!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3943d25c-a9ee-415f-9328-37db9c0549c4_921x340.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!6LvH!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3943d25c-a9ee-415f-9328-37db9c0549c4_921x340.png" width="921" height="340" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/3943d25c-a9ee-415f-9328-37db9c0549c4_921x340.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:340,&quot;width&quot;:921,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!6LvH!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3943d25c-a9ee-415f-9328-37db9c0549c4_921x340.png 424w, https://substackcdn.com/image/fetch/$s_!6LvH!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3943d25c-a9ee-415f-9328-37db9c0549c4_921x340.png 848w, https://substackcdn.com/image/fetch/$s_!6LvH!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3943d25c-a9ee-415f-9328-37db9c0549c4_921x340.png 1272w, https://substackcdn.com/image/fetch/$s_!6LvH!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3943d25c-a9ee-415f-9328-37db9c0549c4_921x340.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div><hr></div><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!-kh9!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!-kh9!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!-kh9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png" width="1456" height="246" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:246,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:172293,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:&quot;&quot;,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!-kh9!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p>I also host an AI podcast and content series called &#8220;Pioneers.&#8221; This series takes you on an enthralling journey into the minds of AI visionaries, founders, and CEOs who are at the forefront of innovation through AI in their organizations.</p><p>To learn more, please visit <a href="https://www.multimodal.dev/blog">Pioneers</a> on Beehiiv.</p><div><hr></div><h2>Wrapping Up</h2><p>If you&#8217;re building or scaling AI-powered solutions, now is the time to assess your requirements, experiment with free tiers, and leverage the strengths of these diverse platforms. The future of AI is not just about picking the most powerful model&#8212;it&#8217;s about finding the right fit for your strategy, workflows, and vision.<br><br>I&#8217;ll come back soon with more such comparisons. <br><br>Until then,<br>Ankur.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.ankursnewsletter.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">If you liked reading this post, consider subscribing for more AI content.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[The Real Price of AI: Pre-Training Vs. Inference Costs]]></title><description><![CDATA[AI inference costs far outweigh training. Learn why they're rising & how to optimize model deployment, hardware & techniques for sustainable AI value.]]></description><link>https://www.ankursnewsletter.com/p/the-real-price-of-ai-pre-training</link><guid isPermaLink="false">https://www.ankursnewsletter.com/p/the-real-price-of-ai-pre-training</guid><dc:creator><![CDATA[Ankur A. Patel]]></dc:creator><pubDate>Thu, 10 Apr 2025 16:43:54 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!1RGK!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F73cf8c4f-95da-4560-9f38-407b10388992_1200x600.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!1RGK!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F73cf8c4f-95da-4560-9f38-407b10388992_1200x600.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!1RGK!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F73cf8c4f-95da-4560-9f38-407b10388992_1200x600.png 424w, https://substackcdn.com/image/fetch/$s_!1RGK!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F73cf8c4f-95da-4560-9f38-407b10388992_1200x600.png 848w, https://substackcdn.com/image/fetch/$s_!1RGK!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F73cf8c4f-95da-4560-9f38-407b10388992_1200x600.png 1272w, https://substackcdn.com/image/fetch/$s_!1RGK!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F73cf8c4f-95da-4560-9f38-407b10388992_1200x600.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!1RGK!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F73cf8c4f-95da-4560-9f38-407b10388992_1200x600.png" width="1200" height="600" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/73cf8c4f-95da-4560-9f38-407b10388992_1200x600.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:600,&quot;width&quot;:1200,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:377433,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.ankursnewsletter.com/i/161032753?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F73cf8c4f-95da-4560-9f38-407b10388992_1200x600.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!1RGK!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F73cf8c4f-95da-4560-9f38-407b10388992_1200x600.png 424w, https://substackcdn.com/image/fetch/$s_!1RGK!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F73cf8c4f-95da-4560-9f38-407b10388992_1200x600.png 848w, https://substackcdn.com/image/fetch/$s_!1RGK!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F73cf8c4f-95da-4560-9f38-407b10388992_1200x600.png 1272w, https://substackcdn.com/image/fetch/$s_!1RGK!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F73cf8c4f-95da-4560-9f38-407b10388992_1200x600.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.ankursnewsletter.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Subscribe today to join our community of 1950 AI enthusiasts and business leaders for more insider takes and practical advice on implementing AI in business.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h2>Key Takeaways</h2><p>Here are five key takeaways from the article:</p><ol><li><p>For most companies using AI, <strong>the ongoing cost of running models daily (inference) vastly outweighs the initial training cost</strong>, potentially accounting for 80-90% of the total lifetime expense.</p></li><li><p>Overall inference costs are rising significantly <strong>due to wider business adoption of AI, the demand for real-time performance, the increasing complexity and size of models, and the growing volume of data processe</strong>d.</p></li><li><p>Proactively managing and <strong>optimizing inference costs is becoming a critical roadblock</strong> to achieving real value and return on investment from generative AI deployments.</p></li><li><p>Strategies to control inference expenses include <strong>choosing the right-sized model for the task, applying optimization techniques like quantization and pruning, making smart hardware choices, and using efficient deployment methods like batching.</strong></p></li><li><p>Successfully leveraging AI <strong>requires a shift towards managing AI systems as continuously operated products</strong>, focusing on optimizing both performance and cost efficiency throughout their lifecycle.</p></li></ol><div><hr></div><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!R9ua!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!R9ua!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!R9ua!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png" width="1456" height="246" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/bea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:246,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:273651,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:&quot;&quot;,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!R9ua!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1456w" sizes="100vw" loading="lazy" fetchpriority="high"></picture><div></div></div></a></figure></div><p>In 2023, I started Multimodal, a Generative AI company that helps organizations automate complex, knowledge-based workflows using AI Agents. Check it out <a href="https://www.multimodal.dev/">here</a>.</p><div><hr></div><p>We're seeing truly rapid advancements in generative AI. It feels like every week brings new capabilities. A lot of the discussion centers on the huge effort and significant amount of hardware needed for training the big large language models, or LLMs.</p><p>But here&#8217;s something critical that often gets missed: for almost any company actually putting AI to work, the day-to-day running &#8211; the inference stage &#8211; tells a very different cost story. Once you have a given model trained, the real cost accumulation begins. Think about 80%, maybe even 90%, of the total dollars spent over a model's active life. This cost is driven by factors like how many users are querying the model, how often, and the volume of input and output tokens processed for each request.</p><p>Let&#8217;s put it plainly: not planning for and managing this inference cost is becoming a major roadblock to getting real value and ROI from AI. In this article, we&#8217;ll talk about inference costs versus pretraining costs, and how to manage them.</p><h2>Demystifying AI Workloads: Training vs. Inference</h2><p>To really grasp the cost dynamics, it helps to clearly separate the two main workloads for generative AI models: training and inference. Think of training as the intense, upfront effort and inference as the ongoing operational work.</p><p>First, let's talk about training. <strong>You can picture training like sending a model to university.</strong> It&#8217;s an intensive process where the goal is to teach the model patterns, structures, and relevant information from vast amounts of data. </p><p>This requires a significant amount of compute power, often using specialized hardware like GPUs or TPUs running for extended periods. It involves crunching through data, adjusting millions or billions of parameters until the model learns effectively. This process results in a large, one-time (or periodic, if retraining) cost factor. It&#8217;s where those headlines about massive hardware demands for large language models originate.</p><p>Now, compare that to inference, or LLM inference specifically. <strong>This is like the graduate actually working their job day-to-day.</strong> The goal here isn't learning; it's applying what was learned during training to new, unseen input data. When users interact with a generative AI application, providing input tokens through prompt engineering or other means, the model performs inference to generate output tokens &#8211; could be text, code, or analysis. </p><p>Each individual inference task might require less compute than the training phase, but here&#8217;s the key: it happens constantly, at scale, for potentially thousands or millions of users. This continuous demand, focused on speed and model performance for a good user experience, is what drives the cumulative inference cost. Achieving efficient inference often requires careful tuning of the software infrastructure.</p><p>This leads to a significant imbalance. For most companies deploying these models, the ongoing inference cost vastly outweighs the initial training cost. <strong>It&#8217;s common for inference to account for <a href="https://rethinkpriorities.org/research-area/gpt-3-like-models-are-now-much-easier-to-access-and-deploy-than-to-develop/">80-90% of the total compute</a> dollars spent over a given model's production lifecycle. </strong>Why? Simply frequency and scale. The model serves far more requests during its operational life than the number of batches processed during its training. This trend makes understanding and reducing inference costs a critical focus for any company looking to deploy AI sustainably.</p><h2>Why Inference Costs are Skyrocketing</h2><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!heai!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F840b482a-f205-47b3-85a7-1715f8d48184_2000x1192.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!heai!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F840b482a-f205-47b3-85a7-1715f8d48184_2000x1192.png 424w, https://substackcdn.com/image/fetch/$s_!heai!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F840b482a-f205-47b3-85a7-1715f8d48184_2000x1192.png 848w, https://substackcdn.com/image/fetch/$s_!heai!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F840b482a-f205-47b3-85a7-1715f8d48184_2000x1192.png 1272w, https://substackcdn.com/image/fetch/$s_!heai!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F840b482a-f205-47b3-85a7-1715f8d48184_2000x1192.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!heai!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F840b482a-f205-47b3-85a7-1715f8d48184_2000x1192.png" width="1456" height="868" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/840b482a-f205-47b3-85a7-1715f8d48184_2000x1192.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:868,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;Unraveling GPU Inference Costs for Fine-tuned Open-source Models V/S Closed  Platforms - MLOps Community&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Unraveling GPU Inference Costs for Fine-tuned Open-source Models V/S Closed  Platforms - MLOps Community" title="Unraveling GPU Inference Costs for Fine-tuned Open-source Models V/S Closed  Platforms - MLOps Community" srcset="https://substackcdn.com/image/fetch/$s_!heai!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F840b482a-f205-47b3-85a7-1715f8d48184_2000x1192.png 424w, https://substackcdn.com/image/fetch/$s_!heai!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F840b482a-f205-47b3-85a7-1715f8d48184_2000x1192.png 848w, https://substackcdn.com/image/fetch/$s_!heai!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F840b482a-f205-47b3-85a7-1715f8d48184_2000x1192.png 1272w, https://substackcdn.com/image/fetch/$s_!heai!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F840b482a-f205-47b3-85a7-1715f8d48184_2000x1192.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><a href="https://uploads-ssl.webflow.com/64128071fa22275256c1c222/648aafa00d85962f5926aef6_image%20(22).png">Source</a></figcaption></figure></div><p>It might seem counterintuitive. We constantly hear about rapid advancements making things cheaper, maybe even hinting at a Moore's Law equivalent for AI cost efficiency. LLM providers often slash the price per million tokens processed. </p><p>Yet, for many organizations, the total bill for running AI models&#8212;the inference cost&#8212;keeps climbing. This isn't a contradiction; it's the result of several powerful trends converging. Let's break down the key factors driving this demand and pushing up the overall dollars spent on inference.</p><h3>Explosion in AI Deployment and Usage</h3><p>The single biggest factor is simply that generative AI is going mainstream within businesses. It's moving rapidly from isolated experiments to being woven into core operations across almost every industry. </p><p>Recent data highlights this adoption curve; for instance, figures cited by <a href="https://newsroom.ibm.com/2024-01-10-Data-Suggests-Growth-in-Enterprise-Adoption-of-AI-is-Due-to-Widespread-Deployment-by-Early-Adopters">TechRadar Pro based on IBM research</a> indicated that around 42% of large companies were already actively using AI, with another 40% actively exploring it. <strong>This wider deployment means more models are running across more business functions&#8212;customer service, marketing, software development, internal analysis, you name it.</strong></p><p>Naturally, more active models serving a larger number of users translates directly into a higher volume of inference calls. Each time an employee uses an AI assistant or a customer interacts with an AI-powered feature, an inference request is made. As usage scales, the aggregate compute demand balloons, driving up the total inference cost for the company, even if the cost per individual interaction is relatively low. This widespread adoption is fundamentally reshaping the market demand for inference capacity.</p><h3>The Need for Speed: Real-Time Performance</h3><p>Many of the most valuable AI applications demand immediate results. Think about conversational AI agents, real-time fraud detection systems, or dynamic recommendation engines. <strong>Users expect instant responses; lag kills the experience and diminishes the value.</strong> Achieving this low latency, or minimizing inference time, often requires significant investment in high-performance hardware infrastructure.</p><p>As discussed in a previous article comparing AI hardware, achieving top-tier speed for complex models often means relying on specialized, powerful accelerators. <strong>This includes NVIDIA's high-end GPUs like the H100 or H200, <a href="https://www.ankursnewsletter.com/p/google-tpus-vs-aws-trainium-and-inferentia">Google's TPUs (especially inference-focused versions like TPU v5e), or AWS's purpose-built Inferentia chips (like Inferentia2)</a>.</strong> </p><p>These hardware options deliver excellent performance but come at a premium price compared to general-purpose CPUs or older GPUs. There's an inherent trade-off: faster speed often means higher infrastructure cost per instance, which directly impacts the inference cost per query. Companies must balance the desired model performance against the budget, making hardware selection a critical factor in managing the final bill from cloud providers like AWS, Google Cloud, or Microsoft Azure.</p><h3>The Complexity Factor: Bigger Models and Deeper Reasoning</h3><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!UnXq!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc4cea6f6-6c72-4a57-a93a-34f1036f1118_1103x720.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!UnXq!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc4cea6f6-6c72-4a57-a93a-34f1036f1118_1103x720.png 424w, https://substackcdn.com/image/fetch/$s_!UnXq!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc4cea6f6-6c72-4a57-a93a-34f1036f1118_1103x720.png 848w, https://substackcdn.com/image/fetch/$s_!UnXq!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc4cea6f6-6c72-4a57-a93a-34f1036f1118_1103x720.png 1272w, https://substackcdn.com/image/fetch/$s_!UnXq!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc4cea6f6-6c72-4a57-a93a-34f1036f1118_1103x720.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!UnXq!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc4cea6f6-6c72-4a57-a93a-34f1036f1118_1103x720.png" width="1103" height="720" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c4cea6f6-6c72-4a57-a93a-34f1036f1118_1103x720.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:720,&quot;width&quot;:1103,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;Navigating the Cost Challenges of Generative AI: Strategies and  Considerations for AI Entrepreneurs&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Navigating the Cost Challenges of Generative AI: Strategies and  Considerations for AI Entrepreneurs" title="Navigating the Cost Challenges of Generative AI: Strategies and  Considerations for AI Entrepreneurs" srcset="https://substackcdn.com/image/fetch/$s_!UnXq!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc4cea6f6-6c72-4a57-a93a-34f1036f1118_1103x720.png 424w, https://substackcdn.com/image/fetch/$s_!UnXq!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc4cea6f6-6c72-4a57-a93a-34f1036f1118_1103x720.png 848w, https://substackcdn.com/image/fetch/$s_!UnXq!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc4cea6f6-6c72-4a57-a93a-34f1036f1118_1103x720.png 1272w, https://substackcdn.com/image/fetch/$s_!UnXq!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc4cea6f6-6c72-4a57-a93a-34f1036f1118_1103x720.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"><a href="https://media.licdn.com/dms/image/v2/D4E12AQFE600cGow8Wg/article-cover_image-shrink_720_1280/article-cover_image-shrink_720_1280/0/1691875054410?e=2147483647&amp;v=beta&amp;t=rm7pTPPEuGnWkEmL-qVjeU3XA7gFTyVVMX5_pD-EnWc">Source</a></figcaption></figure></div><p>Another major driver is the increasing complexity of the models themselves. The state-of-the-art large language models that capture headlines often boast hundreds of billions, or even trillions, of parameters. This sheer model size inherently requires more computation for each inference task compared to smaller models. <strong>Processing the input tokens and generating the required output token sequence for these massive proprietary models or even large open models is computationally intensive.</strong></p><p>Furthermore, we're moving beyond simple pattern recognition or classification towards models capable of sophisticated reasoning. These AI systems are designed to analyze complex situations, break down problems, potentially query external data sources like a vector database for relevant information, evaluate options, and create multi-step plans or detailed analyses. This reasoning process is far more computationally demanding than, say, simple image classification.</p><p>Consider this example: having an AI classify a picture requires one type of inference process. But asking an AI agent to analyze a detailed customer complaint, understand the context, retrieve purchase history, identify relevant policy clauses, compare potential solutions, and then generate a personalized, multi-paragraph resolution plan involves a far more complex sequence. </p><p>Such tasks might even involve the AI agent making multiple internal inference calls to reflect, plan, and refine its output before presenting the final response. This multi-step reasoning significantly multiplies the compute demand&#8212;and thus the inference cost&#8212;for a single user interaction, making these advanced capabilities significantly more expensive to run at scale.</p><h3>The Unrelenting Data Deluge</h3><p>Finally, the ever-increasing volume of data being generated and consumed plays a role. <strong>More data often translates to more input being fed into models for inference, whether it's analyzing larger documents, processing more customer interactions, or monitoring real-time data streams from sensors or financial markets.</strong> This constant flow of input data keeps the inference engines busy, contributing to the overall workload and cost.</p><p>While optimizations like model quantization and hardware improvements work towards reducing inference costs on a per-unit basis, the combined effect of wider adoption,<strong> the demand for real-time speed, the trend towards larger and more complex reasoning models, and growing data volumes is causing the total inference spend for many organizations to rise significantly. </strong></p><p>The AI inference market's projected growth reflects this reality. This makes understanding, measuring, and actively managing inference cost not just a technical challenge, but a critical business imperative for any company looking to achieve sustainable and efficient AI deployment at scale.</p><h2>Strategies for Taming Inference Costs</h2><p>Understanding that inference is the dominant cost factor is the first step. The next, crucial step is actively managing and reducing these costs. Thankfully, there are several effective strategies companies can employ, ranging from model selection to deployment tactics and infrastructure choices. Implementing these can significantly improve the ROI and sustainability of your generative AI initiatives.</p><h3>Right-Sizing Your Models</h3><p>It's tempting to always reach for the largest, most powerful large language models available, whether proprietary models or open models. However, this is often overkill and unnecessarily expensive for many tasks. A core principle of reducing inference costs is model right-sizing.</p><p><strong>Avoid the Sledgehammer</strong>: Don't use a massive, multi-billion parameter model for a task that a much smaller model could handle effectively. Analyze the specific requirements of the application. Does it truly need the nuanced generation capabilities of a giant LLM, or would a task-specific, smaller model suffice for classification, extraction, or simpler Q&amp;A?</p><p><strong>Explore Task-Specific Options</strong>: Often, smaller models trained or fine-tuned for a specific industry or function can deliver excellent performance at a fraction of the inference cost. Sometimes, a "good enough" model performance achieves the business goal much more profitably. Compare the capabilities needed versus the price of running different models.</p><h3>Optimize, Optimize, Optimize</h3><p>Once you've selected a model (or models), significant cost savings can often be achieved through optimization techniques before deployment. These methods aim to make the given model run more efficiently on the hardware, reducing inference time and resource consumption. Key techniques include:</p><ul><li><p><strong>Quantization</strong>: Think of this as reducing the numerical precision of the model's parameters (weights). Instead of using highly precise numbers (like 32-bit floats), quantization might use less precise formats (like 16-bit floats or 8-bit integers). This makes the model smaller and faster to run, often with minimal impact on accuracy for many tasks. It's like using slightly less precise measurements that are much quicker but still accurate enough for the job.</p></li><li><p><strong>Pruning</strong>: This involves identifying and removing redundant or less important connections (parameters) within the neural network, effectively making the model smaller and leaner without significantly degrading its performance. Imagine carefully trimming unnecessary branches from a tree to make it lighter.</p></li><li><p><strong>Distillation</strong>: Here, you train a smaller, more compact model (the "student") to mimic the behavior and output of a larger, more complex model (the "teacher"). The goal is to transfer the knowledge to a more efficient model that's cheaper to run for inference.</p></li></ul><h3>Smart Infrastructure Choices</h3><p>The hardware and cloud infrastructure you use for inference have a massive impact on both speed and cost. Making informed choices here is critical.</p><ul><li><p><strong>Dedicated Inference Hardware</strong>: Specialized chips are designed explicitly for efficient inference. Options like AWS Inferentia2, Google's TPU v5e, and NVIDIA's GPUs (including powerhouses like the H100 and H200, though sometimes overkill if not needed for speed) can offer significantly better performance per watt and performance per dollar compared to general-purpose CPUs for AI workloads. <strong>Selecting the right instance type is key.</strong></p></li><li><p><strong>Leverage Cloud Provider Services</strong>: Major cloud providers (AWS, Google Cloud, Microsoft Azure) offer managed inference services (e.g., SageMaker Inference, Vertex AI Prediction, Azure ML Endpoints). These services often handle scaling, provide optimized software environments, and sometimes integrate model optimization tools, simplifying deployment and potentially lowering costs.</p></li><li><p><strong>Dynamic/Serverless Inference</strong>: For workloads with variable demand, serverless inference platforms can be highly cost-effective. You pay primarily for the compute time consumed during active inference calls, rather than paying for idle instances.</p></li><li><p><strong>Consider Edge Computing</strong>: For applications requiring very low latency or processing sensitive data, running inference directly on edge devices (closer to where data is generated) can sometimes reduce costs associated with data transfer and reliance on centralized cloud infrastructure.</p></li></ul><h3>Efficient Deployment Strategies</h3><p>How you serve inference requests also matters. Simple adjustments can yield efficiency gains.</p><ul><li><p><strong>Batching</strong>: Instead of processing inference requests one by one as they arrive, batching involves grouping multiple requests together and processing them simultaneously. This often improves hardware utilization (especially on GPUs/TPUs) and can significantly increase throughput, lowering the cost per inference.</p></li><li><p><strong>Caching</strong>: For applications where identical input prompts are common, implementing a cache to store and quickly return previous results can avoid redundant model computations, saving both time and cost.</p></li></ul><h3>Continuous Monitoring and Cost Tracking</h3><p>Finally, managing inference cost isn't a one-time setup; it requires ongoing attention.</p><ul><li><p>Implement MLOps: Robust Machine Learning Operations (MLOps) practices are essential. This includes setting up monitoring to track not just model performance and latency, but also the actual inference cost per prediction, per user, or per transaction.</p></li><li><p>Regular Review: Continuously analyze usage patterns and cost data. Are certain features driving disproportionate costs? Can models be further optimized or right-sized based on real-world performance? <strong>This data-driven approach allows for ongoing optimization to keep inference spending aligned with business value.</strong></p></li></ul><div><hr></div><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!-kh9!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!-kh9!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!-kh9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png" width="1456" height="246" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:246,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:172293,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:&quot;&quot;,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!-kh9!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p>I also host an AI podcast and content series called &#8220;Pioneers.&#8221; This series takes you on an enthralling journey into the minds of AI visionaries, founders, and CEOs who are at the forefront of innovation through AI in their organizations.</p><p>To learn more, please visit <a href="https://www.multimodal.dev/blog">Pioneers</a> on Beehiiv.</p><div><hr></div><h2>Wrapping Up</h2><p>As we've explored, the narrative around AI expenses needs a critical update. While training large language models grabs headlines, the persistent, operational cost of LLM inference is rapidly becoming the dominant factor in the total cost of ownership for most companies. Driven by wider adoption, real-time demands, and increasingly complex models, this inference cost demands strategic focus.</p><p>This requires a fundamental mindset shift: moving away from viewing AI implementation as a one-off project and towards managing AI systems as products that require continuous operation and optimization for both performance and cost efficiency. Success isn't just deploying a model; it's running it sustainably at scale.</p><p>Here&#8217;s what I recommend as you navigate this landscape:</p><ul><li><p><strong>Scrutinize Your Initiatives</strong>: Review current and planned AI deployments specifically through the lens of projected inference cost. Don't let it be an afterthought.</p></li><li><p><strong>Integrate Planning</strong>: Foster collaboration between your tech and finance teams to create realistic cost models that cover the entire AI lifecycle, emphasizing operational spending.</p></li><li><p><strong>Prioritize Optimization</strong>: Make reducing inference costs&#8212;through model right-sizing, optimization techniques like quantization, smart hardware and software infrastructure choices&#8212;a core competency within your AI practice.</p></li></ul><p>Ultimately, preparing for and actively managing inference costs isn't just about controlling spending. It's a strategic enabler, crucial for scaling your AI efforts effectively and ensuring you achieve lasting, profitable value from these powerful technologies.</p><p>I&#8217;ll see you next week with more insights on building and deploying enterprise AI.</p><p>Until then,</p><p>Ankur.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.ankursnewsletter.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">If you liked reading this post, consider subscribing for more AI content.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[CoreWeave vs Modal vs Anyscale For Distributed AI Workloads]]></title><description><![CDATA[Compare CoreWeave, Modal, and Anyscale for distributed AI workloads&#8212;exploring performance, scalability, cost efficiency, and best-suited applications.]]></description><link>https://www.ankursnewsletter.com/p/coreweave-vs-modal-vs-anyscale-for</link><guid isPermaLink="false">https://www.ankursnewsletter.com/p/coreweave-vs-modal-vs-anyscale-for</guid><dc:creator><![CDATA[Ankur A. Patel]]></dc:creator><pubDate>Thu, 27 Mar 2025 14:14:39 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!bIWv!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8389cc3f-7163-4e11-a8cc-1aaababb9aad_1200x600.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!bIWv!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8389cc3f-7163-4e11-a8cc-1aaababb9aad_1200x600.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!bIWv!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8389cc3f-7163-4e11-a8cc-1aaababb9aad_1200x600.png 424w, https://substackcdn.com/image/fetch/$s_!bIWv!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8389cc3f-7163-4e11-a8cc-1aaababb9aad_1200x600.png 848w, https://substackcdn.com/image/fetch/$s_!bIWv!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8389cc3f-7163-4e11-a8cc-1aaababb9aad_1200x600.png 1272w, https://substackcdn.com/image/fetch/$s_!bIWv!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8389cc3f-7163-4e11-a8cc-1aaababb9aad_1200x600.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!bIWv!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8389cc3f-7163-4e11-a8cc-1aaababb9aad_1200x600.png" width="1200" height="600" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8389cc3f-7163-4e11-a8cc-1aaababb9aad_1200x600.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:600,&quot;width&quot;:1200,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:388151,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.ankursnewsletter.com/i/159989307?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8389cc3f-7163-4e11-a8cc-1aaababb9aad_1200x600.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!bIWv!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8389cc3f-7163-4e11-a8cc-1aaababb9aad_1200x600.png 424w, https://substackcdn.com/image/fetch/$s_!bIWv!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8389cc3f-7163-4e11-a8cc-1aaababb9aad_1200x600.png 848w, https://substackcdn.com/image/fetch/$s_!bIWv!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8389cc3f-7163-4e11-a8cc-1aaababb9aad_1200x600.png 1272w, https://substackcdn.com/image/fetch/$s_!bIWv!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8389cc3f-7163-4e11-a8cc-1aaababb9aad_1200x600.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.ankursnewsletter.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Subscribe today to join our community of 1920 AI enthusiasts and business leaders for more insider takes and practical advice on implementing AI in business.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h2><strong>Key Takeaways</strong></h2><ol><li><p><strong>CoreWeave</strong> offers a Kubernetes-native, high-performance infrastructure with NVIDIA GPUs and InfiniBand networking, excelling in AI model training, VFX rendering, and simulations.</p></li><li><p><strong>Modal</strong> simplifies serverless AI development with dynamic autoscaling, Python-based workflows, and second-by-second billing, ideal for flexible GPU workloads and real-time applications.</p></li><li><p><strong>Anyscale</strong>, powered by Ray, provides robust distributed computing with RayTurbo optimization, seamless hybrid cloud integration, and advanced orchestration tools like Apache Airflow.</p></li><li><p>Each platform targets specific AI workload demands: <strong>CoreWeave for sustained high-throughput tasks, Modal for bursty workloads, and Anyscale for complex distributed experiments.</strong></p></li><li><p>Choosing the right platform depends on <strong>workload complexity, scalability needs, cost considerations, and integration with existing tools</strong> and frameworks.</p></li></ol><div><hr></div><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!R9ua!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!R9ua!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!R9ua!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png" width="1456" height="246" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/bea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:246,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:273651,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:&quot;&quot;,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!R9ua!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1456w" sizes="100vw" loading="lazy" fetchpriority="high"></picture><div></div></div></a></figure></div><p>In 2023, I started Multimodal, a Generative AI company that helps organizations automate complex, knowledge-based workflows using AI Agents. Check it out <a href="https://www.multimodal.dev/">here</a>.</p><div><hr></div><p>Today, there is an unprecedented demand for distributed AI workloads as businesses seek to process vast datasets and deploy advanced models at scale.</p><p>From generative AI to real-time inference, these workloads require immense computational power, low latency, and seamless scalability. A recent survey revealed that <a href="https://broadbandbreakfast.com/ciena-survey-predicts-surge-in-ai-driven-data-center-demands/">43% of new data centers are now dedicated to AI</a>. Choosing the right platform is critical&#8212;not only to optimize performance and costs but also to ensure compatibility with specific workload requirements and long-term scalability.</p><p>Let&#8217;s compare CoreWeave, Anyscale, and Modal for distributed AI workloads and figure out which applications each works best for.</p><h2>CoreWeave</h2><h3>Technical Highlights</h3><p><a href="https://www.ankursnewsletter.com/p/comparing-ai-cloud-providers-coreweave?utm_source=publication-search">CoreWeave</a> is a <strong>Kubernetes-native cloud provider</strong> purpose-built for AI workloads, offering unmatched performance and flexibility. Its infrastructure is optimized for <strong>high-performance compute</strong>, enabling users to train and deploy AI models with speed and efficiency. Key features include:</p><ul><li><p><strong>Broad GPU Options</strong>: Access to NVIDIA&#8217;s latest GPUs, including H100 Tensor Core, ensures support for demanding AI applications like generative AI and large-scale simulations.</p></li><li><p><strong>Rapid Resource Provisioning</strong>: Spin up instances in as little as 5 seconds, enabling on-demand scalability for developers and teams.</p></li><li><p><strong>Specialized Networking</strong>: Powered by NVIDIA Quantum InfiniBand, CoreWeave delivers ultra-low latency and up to 400 GB/s throughput, critical for distributed training at scale.</p></li><li><p><strong>AI Object Storage</strong>: High-speed storage provides up to 2 GB/s per GPU, minimizing bottlenecks during data-intensive training or inference tasks.</p></li></ul><h3>Best Suited Applications</h3><p>CoreWeave excels in scenarios requiring massive scale and compute-intensive workloads. Its <strong>unified AI platform</strong> supports:</p><ul><li><p><strong>Large-scale AI model training and fine-tuning</strong>: Ideal for foundational model development or hyperparameter tuning.</p></li><li><p><strong>Visual Effects (VFX) and Rendering</strong>: Accelerates artist workflows with real-time rendering capabilities.</p></li><li><p><strong>Complex Simulations</strong>: Supports industries like financial analytics and life sciences with robust infrastructure for simulations.</p></li><li><p><strong>Generative AI Workloads</strong>: Perfect for cutting-edge applications like LLMs or multimodal AI models requiring high scalability and reliability.</p></li></ul><h2>Modal</h2><h3>Technical Highlights</h3><p>Modal redefines serverless computing by empowering developers to focus on code while it manages the infrastructure behind the scenes. Designed for flexibility, its features include:</p><ul><li><p><strong>Serverless Compute Platform</strong>: Run Python functions at scale without worrying about infrastructure complexity.</p></li><li><p><strong>Autoscaling Capabilities</strong>: Dynamically adjusts resources based on workload demand, ensuring cost efficiency and high machine utilization.</p></li><li><p><strong>High Resource Limits per Container</strong>: Supports up to 64 CPUs, 336 GB of memory, and 8 NVIDIA H100 GPUs, making it ideal for heavy-duty AI tasks like training or inference pipelines.</p></li><li><p><strong>Second-by-second Billing</strong>: Scales to zero when idle, ensuring developers only pay for what they use.</p></li></ul><h3>Best Suited Applications</h3><p>Modal is tailored for developers seeking simplicity without sacrificing power or scalability. It shines in:</p><ul><li><p><strong>Data-intensive Video Processing</strong>: Efficiently handles large-scale video analysis or transformation tasks.</p></li><li><p><strong>Custom ETL Jobs and Periodic Tasks</strong>: Automates data pipelines with scheduled Python functions.</p></li><li><p><strong>Flexible GPU Workloads</strong>: Supports diverse AI applications requiring GPU acceleration.</p></li><li><p><strong>Real-time Production AI Workloads</strong>: Ideal for deploying web services or APIs that need consistent performance under varying loads.</p></li></ul><h2>Anyscale</h2><h3>Technical Highlights</h3><p>Built on the open-source Ray framework, Anyscale transforms distributed computing with a focus on ease-of-use, scalability, and developer productivity. Highlights include:</p><ul><li><p><strong>RayTurbo Optimization</strong>: Enhances Ray&#8217;s performance with improved reliability, cost efficiency, and fault tolerance at scale.</p></li><li><p><strong>Unified Development Environment</strong>: Anyscale Workspaces integrate seamlessly with tools like Jupyter Notebooks and VS Code, enabling developers to build, test, and deploy without context switching.</p></li><li><p><strong>Customizable Deployments</strong>: Supports public clouds (AWS/GCP), on-premises setups, or Kubernetes clusters, giving users complete control over their infrastructure.</p></li><li><p><strong>Built-in Observability Tools</strong>: Provides real-time monitoring and job scheduling to simplify management of distributed workloads.</p></li></ul><h3>Best Suited Applications</h3><p>Anyscale is ideal for teams managing complex workflows across development and production environments. Its strengths include:</p><ul><li><p><strong>Distributed Data Processing with Ray Data</strong>: Scales Python-based pipelines effortlessly across nodes for efficient data handling.</p></li><li><p><strong>Large-scale Model Training &amp; Hyperparameter Tuning</strong>: Optimized for iterative experimentation at massive scale.</p></li><li><p><strong>Serving Complex ML Applications</strong>: Handles high-volume requests with low latency using Ray Serve&#8217;s advanced capabilities.</p></li><li><p><strong>Generative AI Workloads (LLM Fine-tuning)</strong>: Supports fine-tuning large language models or retrieval-augmented generation (RAG) pipelines with precision orchestration tools like RayTurbo.</p></li></ul><h2>Comparison</h2><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!MGpb!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe61d0fa3-9bf4-4a45-846f-c3cf4104b525_1340x415.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!MGpb!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe61d0fa3-9bf4-4a45-846f-c3cf4104b525_1340x415.png 424w, https://substackcdn.com/image/fetch/$s_!MGpb!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe61d0fa3-9bf4-4a45-846f-c3cf4104b525_1340x415.png 848w, https://substackcdn.com/image/fetch/$s_!MGpb!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe61d0fa3-9bf4-4a45-846f-c3cf4104b525_1340x415.png 1272w, https://substackcdn.com/image/fetch/$s_!MGpb!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe61d0fa3-9bf4-4a45-846f-c3cf4104b525_1340x415.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!MGpb!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe61d0fa3-9bf4-4a45-846f-c3cf4104b525_1340x415.png" width="1340" height="415" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e61d0fa3-9bf4-4a45-846f-c3cf4104b525_1340x415.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:415,&quot;width&quot;:1340,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!MGpb!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe61d0fa3-9bf4-4a45-846f-c3cf4104b525_1340x415.png 424w, https://substackcdn.com/image/fetch/$s_!MGpb!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe61d0fa3-9bf4-4a45-846f-c3cf4104b525_1340x415.png 848w, https://substackcdn.com/image/fetch/$s_!MGpb!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe61d0fa3-9bf4-4a45-846f-c3cf4104b525_1340x415.png 1272w, https://substackcdn.com/image/fetch/$s_!MGpb!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe61d0fa3-9bf4-4a45-846f-c3cf4104b525_1340x415.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h3>Infrastructure Flexibility</h3><p>When it comes to infrastructure flexibility, <strong>CoreWeave AI</strong>, <strong>Modal</strong>, and <strong>Anyscale</strong> each offer distinct advantages tailored to specific AI workloads.</p><p>CoreWeave excels with its <strong>Kubernetes-native architecture</strong>, enabling seamless orchestration of GPU resources across bare-metal nodes. This design is ideal for distributed training frameworks like PyTorch Elastic and Horovod, providing unparalleled control over resource allocation and scaling.</p><p>Modal, on the other hand, simplifies infrastructure management with its <strong>serverless platform</strong>, allowing developers to execute <strong>Python functions</strong> without worrying about provisioning or lifecycle management.</p><p>Anyscale stands out by leveraging Ray&#8217;s distributed computing capabilities to integrate seamlessly across public clouds, private clouds, or hybrid environments, offering unmatched flexibility for scaling <strong>AI models</strong> and workloads anywhere.</p><h3>Performance and Scalability</h3><p>For high-performance compute at scale, CoreWeave leads with its <strong>optimized GPU clusters featuring NVIDIA H100 GPUs and Quantum InfiniBand networking</strong>. These features significantly reduce communication overhead during distributed training, achieving up to 20% higher performance compared to general-purpose cloud providers.</p><p>Modal focuses on <strong>rapid scalability, enabling sub-second container starts and scaling to hundreds of GPUs in seconds</strong>&#8212;ideal for bursty workloads like generative AI inference or batch processing.</p><p>Anyscale leverages RayTurbo for <strong>enhanced autoscaling performance, launching clusters of thousands of nodes in under a minute while dynamically optimizing resource utilization.</strong></p><p>Each platform enables massive scale but targets different workload demands: CoreWeave for sustained high-throughput tasks, Modal for flexible scaling, and Anyscale for distributed experiments and production pipelines.</p><h3>Developer Experience</h3><p>CoreWeave&#8217;s Kubernetes Service (CKS) <strong>provides developers with direct access to GPU-accelerated infrastructure while integrating tools like Slurm for efficient workload scheduling.</strong></p><p>Modal simplifies the <strong>developer experience by abstracting away infrastructure complexities, allowing users to focus solely on writing code via Python decorators or pre-configured environments.</strong></p><p>Anyscale takes developer experience further with its unified Workspaces environment, <strong>enabling seamless transitions from local development to cloud-scale deployments without modifying code.</strong></p><p>All three platforms integrate popular frameworks such as TensorFlow and PyTorch, but Anyscale&#8217;s integration with Apache Airflow adds advanced workflow orchestration capabilities for managing complex pipelines.</p><h3>Cost Optimization</h3><p>Cost optimization is a priority across all platforms but is approached differently. CoreWeave achieves efficiency through purpose-built infrastructure that minimizes idle costs while maximizing GPU utilization during training and inference.</p><p>Modal employs a serverless pricing model that bills users by the second, ensuring cost efficiency for bursty workloads or batch processing tasks.</p><p>Anyscale leverages spot instances and dynamic cluster resizing to lower costs while maintaining reliability during peak demand. Additionally, Anyscale&#8217;s ability to optimize resource allocation across heterogeneous instance types (GPUs, TPUs) ensures customers only pay for what they need&#8212;making it particularly attractive for organizations managing diverse workloads.</p><h3>Integration with Existing Tools</h3><p>CoreWeave integrates seamlessly with Kubernetes-based workflows and distributed training frameworks like PyTorch Elastic, making it ideal for teams already leveraging containerized environments.</p><p><strong>Modal supports direct deployment of custom models alongside popular frameworks like Stable Diffusion or GPT-J, offering pre-configured environments that work out-of-the-box.</strong></p><p>Anyscale shines in its integration capabilities by embedding Ray into tools like Apache Airflow, enabling teams to orchestrate distributed tasks within familiar ecosystems while benefiting from RayTurbo&#8217;s performance optimizations.</p><p>These integrations reduce complexity and streamline workflows across development and production environments.</p><div><hr></div><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!-kh9!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!-kh9!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!-kh9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png" width="1456" height="246" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:246,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:172293,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:&quot;&quot;,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!-kh9!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p>I also host an AI podcast and content series called &#8220;Pioneers.&#8221; This series takes you on an enthralling journey into the minds of AI visionaries, founders, and CEOs who are at the forefront of innovation through AI in their organizations.</p><p>To learn more, please visit <a href="https://www.multimodal.dev/blog">Pioneers</a> on Beehiiv.</p><div><hr></div><h2><strong>Wrap Up</strong></h2><p>While all three platforms excel in supporting AI workloads at scale, the choice ultimately depends on specific requirements:</p><ul><li><p><strong>CoreWeave AI</strong> is best suited for high-throughput training and inference tasks requiring specialized infrastructure optimized for efficiency and performance.</p></li><li><p><strong>Modal</strong> offers unmatched simplicity and flexibility for serverless applications like generative AI inference or batch processing.</p></li><li><p><strong>Anyscale</strong> provides robust distributed computing capabilities ideal for teams managing complex workflows or scaling experiments across diverse infrastructures.</p></li></ul><p>I&#8217;ll come back next week with more such insights and comparisons. <br><br>Until then,</p><p>Ankur.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.ankursnewsletter.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">If you liked reading this post, consider subscribing for more AI content.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Reasoning Vs Non-Reasoning LLMs: Architectural Tradeoffs]]></title><description><![CDATA[Explore the architectural divide in modern AI: specialized reasoning engines vs. general-purpose models, and the emerging standards shaping their deployment.]]></description><link>https://www.ankursnewsletter.com/p/reasoning-vs-non-reasoning-llms-architectural</link><guid isPermaLink="false">https://www.ankursnewsletter.com/p/reasoning-vs-non-reasoning-llms-architectural</guid><dc:creator><![CDATA[Ankur A. Patel]]></dc:creator><pubDate>Thu, 20 Mar 2025 17:04:23 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!i77u!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5f5849b6-a456-4c9e-8a74-fd863fe08c4f_1200x600.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!i77u!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5f5849b6-a456-4c9e-8a74-fd863fe08c4f_1200x600.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!i77u!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5f5849b6-a456-4c9e-8a74-fd863fe08c4f_1200x600.png 424w, https://substackcdn.com/image/fetch/$s_!i77u!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5f5849b6-a456-4c9e-8a74-fd863fe08c4f_1200x600.png 848w, https://substackcdn.com/image/fetch/$s_!i77u!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5f5849b6-a456-4c9e-8a74-fd863fe08c4f_1200x600.png 1272w, https://substackcdn.com/image/fetch/$s_!i77u!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5f5849b6-a456-4c9e-8a74-fd863fe08c4f_1200x600.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!i77u!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5f5849b6-a456-4c9e-8a74-fd863fe08c4f_1200x600.png" width="1200" height="600" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/5f5849b6-a456-4c9e-8a74-fd863fe08c4f_1200x600.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:600,&quot;width&quot;:1200,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:385549,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.ankursnewsletter.com/i/159495115?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5f5849b6-a456-4c9e-8a74-fd863fe08c4f_1200x600.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!i77u!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5f5849b6-a456-4c9e-8a74-fd863fe08c4f_1200x600.png 424w, https://substackcdn.com/image/fetch/$s_!i77u!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5f5849b6-a456-4c9e-8a74-fd863fe08c4f_1200x600.png 848w, https://substackcdn.com/image/fetch/$s_!i77u!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5f5849b6-a456-4c9e-8a74-fd863fe08c4f_1200x600.png 1272w, https://substackcdn.com/image/fetch/$s_!i77u!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5f5849b6-a456-4c9e-8a74-fd863fe08c4f_1200x600.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.ankursnewsletter.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Subscribe today to join our community of 1901 AI enthusiasts and business leaders for more insider takes and practical advice on implementing AI in business.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h2>Key Takeaways</h2><ol><li><p>MoE architectures like DeepSeek R1 <strong>offer significant efficiency gains</strong> over dense models.</p></li><li><p>Grok-3's "Big Brain" mode demonstrates the power of <strong>massive computational resources for complex reasoning</strong>.</p></li><li><p>Hybrid models like Claude 3.7 Sonnet <strong>provide flexible reasoning capabilities</strong> with controlled compute allocation.</p></li><li><p>Emerging deployment standards are defining clear use cases for <strong>reasoning vs. pattern-matching models.</strong></p></li><li><p>The future of AI deployment lies in <strong>strategic combinations of specialized and general-purpose architectures.</strong></p></li></ol><div><hr></div><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!R9ua!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!R9ua!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!R9ua!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png" width="1456" height="246" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/bea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:246,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:273651,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:&quot;&quot;,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!R9ua!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1456w" sizes="100vw" loading="lazy" fetchpriority="high"></picture><div></div></div></a></figure></div><p>In 2023, I started Multimodal, a Generative AI company that helps organizations automate complex, knowledge-based workflows using AI Agents. Check it out <a href="https://www.multimodal.dev/">here</a>.</p><div><hr></div><p>Today, a clear divide is emerging between specialized reasoning engines and general-purpose hybrid models. Let&#8217;s dive into the architectural tradeoffs that define modern AI systems, exploring how different approaches tackle the challenges of scale, efficiency, and versatility.</p><h2>Architectural Paradigms in Modern LLMs</h2><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!yo4k!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0ae94873-84a6-40e4-9bb8-0554794ab706_1480x561.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!yo4k!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0ae94873-84a6-40e4-9bb8-0554794ab706_1480x561.png 424w, https://substackcdn.com/image/fetch/$s_!yo4k!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0ae94873-84a6-40e4-9bb8-0554794ab706_1480x561.png 848w, https://substackcdn.com/image/fetch/$s_!yo4k!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0ae94873-84a6-40e4-9bb8-0554794ab706_1480x561.png 1272w, https://substackcdn.com/image/fetch/$s_!yo4k!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0ae94873-84a6-40e4-9bb8-0554794ab706_1480x561.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!yo4k!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0ae94873-84a6-40e4-9bb8-0554794ab706_1480x561.png" width="1456" height="552" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/0ae94873-84a6-40e4-9bb8-0554794ab706_1480x561.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:552,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:64406,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.ankursnewsletter.com/i/159495115?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0ae94873-84a6-40e4-9bb8-0554794ab706_1480x561.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!yo4k!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0ae94873-84a6-40e4-9bb8-0554794ab706_1480x561.png 424w, https://substackcdn.com/image/fetch/$s_!yo4k!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0ae94873-84a6-40e4-9bb8-0554794ab706_1480x561.png 848w, https://substackcdn.com/image/fetch/$s_!yo4k!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0ae94873-84a6-40e4-9bb8-0554794ab706_1480x561.png 1272w, https://substackcdn.com/image/fetch/$s_!yo4k!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0ae94873-84a6-40e4-9bb8-0554794ab706_1480x561.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h3>Specialized Reasoning Engines</h3><h4>DeepSeek R1 (MoE Architecture)</h4><p>DeepSeek R1 represents a paradigm shift in AI model design, leveraging a Mixture of Experts (MoE) architecture to achieve unprecedented efficiency. At its core, R1 activates only 37 billion of its 671 billion parameters per token, enabling sparse computation that dramatically reduces computational overhead. <strong>This approach allows R1 to maintain a massive knowledge base while operating with the agility of much smaller models.</strong></p><p>R1's chain-of-thought reinforcement learning optimization incorporates self-correction mechanisms, enhancing its ability to reason through complex problems and refine its outputs iteratively. With a 64K context window and a dedicated 32K token budget for chain-of-thought reasoning, R1 can handle extended dialogues and multi-step problem-solving with ease.</p><p>Perhaps most notably, DeepSeek R1's MIT-licensed MoE implementation achieves an 82% reduction in floating-point operations (FLOPs) compared to dense models of similar scale. This efficiency gain translates directly to lower inference costs and faster response times, making R1 a compelling option for resource-conscious deployments.</p><h4>Grok-3 ("Big Brain" Mode)</h4><p>Elon Musk's xAI has taken a different approach with Grok-3, focusing on raw computational power and advanced reasoning capabilities. <strong>Grok-3's "Big Brain" mode harnesses a staggering 100,000 Nvidia H100 GPU cluster, enabling it to tackle computationally intensive tasks with unparalleled speed and depth.</strong></p><p>At the heart of Grok-3's architecture is a multi-agent debate system that validates solutions through simulated discourse between AI agents. This approach mimics human problem-solving dynamics, allowing Grok-3 to explore multiple perspectives and arrive at more robust conclusions.</p><p>Grok-3's dynamic compute scaling adjusts processing resources on the fly, delivering responses in as little as 1-10 seconds for even the most complex queries. This flexibility ensures that users receive timely responses without sacrificing depth of analysis.</p><p>The integration of DeepSearch technology gives Grok-3 the ability to verify information in real-time by scanning the internet and X (formerly Twitter) for the latest data. This feature keeps Grok-3's knowledge base current and enhances its ability to provide accurate, up-to-date information.</p><h3>General-Purpose Hybrid Models</h3><h4>Claude 3.7 Sonnet</h4><p>Anthropic's Claude 3.7 Sonnet takes a hybrid approach, blending the versatility of general-purpose models with advanced reasoning capabilities. Built upon the foundation of Claude 3.5, Sonnet incorporates reinforcement learning from human feedback (RLHF) to refine its outputs and align more closely with human expectations.</p><p><strong>A standout feature of Claude 3.7 is its token-controlled reasoning, allowing users to allocate between 0 and 128,000 tokens for "thinking" before generating a response.</strong> This granular control over the model's cognitive resources enables a balance between quick replies and deep, thoughtful analysis as needed.</p><p>Sonnet boasts a 200K token context window with modified self-attention mechanisms, allowing it to maintain coherence over extremely long conversations or documents. Despite its expanded capabilities, Claude 3.7 achieves a 37% faster throughput compared to GPT-4.5, highlighting Anthropic's focus on performance optimization.</p><h4>GPT-4.5</h4><p>OpenAI's GPT-4.5 represents the pinnacle of dense model architectures, featuring a massive 12.8 trillion parameter count. This sheer scale allows GPT-4.5 to capture intricate patterns and relationships across a vast range of topics and tasks.</p><p><strong>A key focus in GPT-4.5's development has been enhancing its emotional intelligence, enabling more nuanced and empathetic interactions.</strong> This advancement positions GPT-4.5 as a powerful tool for applications requiring deep understanding of human sentiment and social dynamics.</p><p>With a 128K token context window, GPT-4.5 can handle extended conversations and analyze lengthy documents. However, this capability comes at a premium, with output costs reaching $150 per million tokens. This pricing structure reflects the significant computational resources required to operate such a large, dense model at scale.</p><h2>Emerging Deployment Standards</h2><p>As AI technologies mature, clear patterns are emerging in how different types of models are deployed for specific use cases. This section explores the emerging standards for deploying reasoning models, pattern-matching models, and hybrid approaches.</p><h3>Reasoning Model Applications</h3><p>Reasoning models excel in domains requiring complex, multi-step analysis and decision-making. Two key areas where reasoning models are becoming the standard are medical diagnosis and supply chain optimization.</p><h4>Medical Diagnosis</h4><p>In the medical field, reasoning models are help with diagnostic processes:</p><ul><li><p><strong>Multi-step differential diagnosis flows:</strong> These models can navigate through complex decision trees, considering a wide range of symptoms, test results, and patient history to arrive at accurate diagnoses.</p></li><li><p><strong>Probabilistic reasoning over symptom clusters:</strong> By analyzing the likelihood of various conditions based on combinations of symptoms, these models can provide more nuanced and accurate diagnoses, especially for rare or complex cases.</p></li></ul><h2><strong>Supply Chain Optimization</strong></h2><p>Reasoning models are also transforming supply chain management:</p><ul><li><p><strong>Constraint-based reasoning for routing:</strong> These models can optimize complex logistics networks by considering multiple constraints such as delivery times, vehicle capacities, and fuel efficiency.</p></li><li><p><strong>Dynamic replanning with Monte Carlo Tree Search:</strong> This approach allows for real-time adjustments to supply chain strategies, adapting to unexpected events or changes in demand.</p></li></ul><h2>Pattern-Matching Dominance Areas</h2><p>While reasoning models excel in complex decision-making, pattern-matching models remain superior in areas requiring rapid processing of large datasets and identification of subtle patterns.</p><h3>Content Recommendations</h3><p>In the realm of content recommendations, pattern-matching models continue to dominate:</p><ul><li><p><strong>Attention-based sequence modeling:</strong> These models excel at understanding user preferences by analyzing sequences of interactions, leading to more personalized recommendations.</p></li><li><p><strong>Real-time collaborative filtering:</strong> By quickly identifying patterns across large user bases, these models can provide up-to-the-minute recommendations based on current trends and user behaviors.</p></li></ul><h2>Classification Tasks</h2><p>For classification tasks, pattern-matching models offer unparalleled efficiency:</p><ul><li><p><strong>Fine-tuned LoRA adapters:</strong> These lightweight adaptations allow for rapid customization of large language models for specific classification tasks, improving both speed and accuracy.</p></li><li><p><strong>Ensemble pattern detectors:</strong> By combining multiple specialized classifiers, these models can achieve high accuracy across a wide range of classification tasks.</p></li></ul><h2>Hybrid Deployment Strategies</h2><p>As the lines between reasoning and pattern-matching blur, hybrid deployment strategies are emerging to leverage the strengths of both approaches.</p><h3>Claude 3.7's Adaptive Compute Budgeting</h3><p>Anthropic's Claude 3.7 introduces a novel approach to balancing quick responses with deep reasoning:</p><ul><li><p><strong>API-controlled thinking tokens:</strong> Users can allocate between $3 and $15 per million tokens for "thinking," allowing fine-grained control over the depth of reasoning applied to a given task.</p></li><li><p><strong>Early exit mechanisms:</strong> For simpler queries, the model can provide quick responses without engaging in unnecessary deep reasoning, optimizing both speed and resource usage.</p></li></ul><h3>Grok-3's Mode Switching</h3><p>xAI's Grok-3 takes a different approach to hybrid deployment:</p><ul><li><p><strong>Automatic Big Brain activation:</strong> The model can automatically switch to a more intensive reasoning mode for complex tasks, including:</p><ul><li><p>Multi-variable calculus problems</p></li><li><p>Legal document analysis</p></li><li><p>Counterfactual reasoning scenarios</p></li></ul></li></ul><p>This adaptive approach ensures that the appropriate level of computational resources is applied based on the complexity of the task at hand.</p><h2><strong>Architectural Tradeoff Matrix</strong></h2><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!8hOX!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5906a64c-c5c6-492e-af6f-8bd0862531e4_651x227.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!8hOX!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5906a64c-c5c6-492e-af6f-8bd0862531e4_651x227.png 424w, https://substackcdn.com/image/fetch/$s_!8hOX!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5906a64c-c5c6-492e-af6f-8bd0862531e4_651x227.png 848w, https://substackcdn.com/image/fetch/$s_!8hOX!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5906a64c-c5c6-492e-af6f-8bd0862531e4_651x227.png 1272w, https://substackcdn.com/image/fetch/$s_!8hOX!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5906a64c-c5c6-492e-af6f-8bd0862531e4_651x227.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!8hOX!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5906a64c-c5c6-492e-af6f-8bd0862531e4_651x227.png" width="651" height="227" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/5906a64c-c5c6-492e-af6f-8bd0862531e4_651x227.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:227,&quot;width&quot;:651,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!8hOX!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5906a64c-c5c6-492e-af6f-8bd0862531e4_651x227.png 424w, https://substackcdn.com/image/fetch/$s_!8hOX!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5906a64c-c5c6-492e-af6f-8bd0862531e4_651x227.png 848w, https://substackcdn.com/image/fetch/$s_!8hOX!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5906a64c-c5c6-492e-af6f-8bd0862531e4_651x227.png 1272w, https://substackcdn.com/image/fetch/$s_!8hOX!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5906a64c-c5c6-492e-af6f-8bd0862531e4_651x227.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><div><hr></div><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!-kh9!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!-kh9!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!-kh9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png" width="1456" height="246" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:246,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:172293,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:&quot;&quot;,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!-kh9!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p>I also host an AI podcast and content series called &#8220;Pioneers.&#8221; This series takes you on an enthralling journey into the minds of AI visionaries, founders, and CEOs who are at the forefront of innovation through AI in their organizations.</p><p>To learn more, please visit <a href="https://www.multimodal.dev/blog">Pioneers</a> on Beehiiv.</p><div><hr></div><h2>Wrapping Up</h2><p>Each approach offers unique strengths: MoE architectures provide efficiency, while dense models offer versatility. Hybrid solutions are emerging to bridge this gap, offering adaptive compute and specialized modes.</p><p>As deployment standards crystallize, we see reasoning models excelling in complex decision-making tasks, while pattern-matching models dominate in rapid data processing. The future of AI lies in strategically leveraging these diverse architectures to optimize performance, cost-efficiency, and applicability across a wide range of industries and use cases.</p><p>I&#8217;ll come back next week with more on model choice and tradeoffs. <br><br>Until then,</p><p>Ankur.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.ankursnewsletter.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">If you liked reading this post, consider subscribing for more AI content.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Pinecone vs. Weaviate vs. Milvus for Vector Search]]></title><description><![CDATA[Compare Pinecone, Weaviate, and Milvus&#8212;leading vector databases for AI applications. Explore their strengths in vector search, scalability, and deployment.]]></description><link>https://www.ankursnewsletter.com/p/pinecone-vs-weaviate-vs-milvus-for</link><guid isPermaLink="false">https://www.ankursnewsletter.com/p/pinecone-vs-weaviate-vs-milvus-for</guid><dc:creator><![CDATA[Ankur A. Patel]]></dc:creator><pubDate>Thu, 06 Mar 2025 16:44:35 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!L1Vy!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff853c535-0af7-48e6-acef-c445d453fa55_1200x600.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!L1Vy!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff853c535-0af7-48e6-acef-c445d453fa55_1200x600.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!L1Vy!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff853c535-0af7-48e6-acef-c445d453fa55_1200x600.png 424w, https://substackcdn.com/image/fetch/$s_!L1Vy!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff853c535-0af7-48e6-acef-c445d453fa55_1200x600.png 848w, https://substackcdn.com/image/fetch/$s_!L1Vy!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff853c535-0af7-48e6-acef-c445d453fa55_1200x600.png 1272w, https://substackcdn.com/image/fetch/$s_!L1Vy!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff853c535-0af7-48e6-acef-c445d453fa55_1200x600.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!L1Vy!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff853c535-0af7-48e6-acef-c445d453fa55_1200x600.png" width="1200" height="600" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f853c535-0af7-48e6-acef-c445d453fa55_1200x600.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:600,&quot;width&quot;:1200,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:379395,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.ankursnewsletter.com/i/158525191?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff853c535-0af7-48e6-acef-c445d453fa55_1200x600.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!L1Vy!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff853c535-0af7-48e6-acef-c445d453fa55_1200x600.png 424w, https://substackcdn.com/image/fetch/$s_!L1Vy!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff853c535-0af7-48e6-acef-c445d453fa55_1200x600.png 848w, https://substackcdn.com/image/fetch/$s_!L1Vy!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff853c535-0af7-48e6-acef-c445d453fa55_1200x600.png 1272w, https://substackcdn.com/image/fetch/$s_!L1Vy!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff853c535-0af7-48e6-acef-c445d453fa55_1200x600.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.ankursnewsletter.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Subscribe today to join our community of 1890 AI enthusiasts and business leaders for more insider takes and practical advice on implementing AI in business.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h2><strong>Key Takeaways</strong></h2><ol><li><p>Vector databases like <strong>Pinecone, Weaviate, and Milvus</strong> play a pivotal role in AI applications by efficiently handling high-dimensional vector embeddings for tasks such as semantic search, recommendation systems, and generative AI.</p></li><li><p>Pinecone excels in <strong>real-time vector similarity searches</strong> with minimal setup and low-latency performance, making it ideal for applications requiring fast query responses and cloud-native infrastructure.</p></li><li><p>Weaviate's <strong>hybrid search capabilities and support</strong> for multi-modal data bridge the gap between structured and unstructured data, providing flexibility for projects involving semantic meaning and complex queries.</p></li><li><p>Milvus stands out for its <strong>scalability and modular architecture</strong>, enabling efficient similarity searches across massive datasets and diverse indexing methods tailored to specific project requirements.</p></li><li><p>Choosing the right vector database depends on <strong>deployment preferences, scalability needs, query complexity, and cost considerations</strong>, with Pinecone suited for ease of use, Weaviate for hybrid search, and Milvus for large-scale AI workloads.</p><div><hr></div></li></ol><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!R9ua!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!R9ua!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!R9ua!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png" width="1456" height="246" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/bea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:246,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:273651,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:&quot;&quot;,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!R9ua!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1456w" sizes="100vw" loading="lazy" fetchpriority="high"></picture><div></div></div></a></figure></div><p>In 2023, I started Multimodal, a Generative AI company that helps organizations automate complex, knowledge-based workflows using AI Agents. Check it out <a href="https://www.multimodal.dev/">here</a>.</p><div><hr></div><p>Vector databases play a pivotal role in modern AI applications, enabling efficient handling of high-dimensional vector embeddings generated by machine learning models. <strong>Unlike traditional relational databases, they excel in managing vast amounts of unstructured data, supporting semantic search, recommendation systems, and generative AI tasks</strong>. Pinecone, Weaviate, and Milvus are leading solutions in this space, each offering unique capabilities for vector similarity searches and complex queries.</p><p>Let&#8217;s explore their strengths in handling structured and unstructured data, scalability for massive datasets, and deployment options to help developers choose the ideal vector database for specific project requirements.</p><h2><strong>Overview of Each Database</strong></h2><h2><strong>Pinecone</strong></h2><p>Pinecone is a fully managed, cloud-native vector database designed to simplify vector search and deliver real-time processing for high-dimensional vectors. It excels in handling vast amounts of vector data with minimal setup, making it a go-to choice for AI applications requiring low-latency performance and seamless scalability.</p><h4><strong>Key Features</strong></h4><ul><li><p><strong>Cloud-native infrastructure:</strong> Eliminates the complexity of infrastructure management, allowing users to focus on generating embeddings and retrieving relevant results.</p></li><li><p><strong>Real-time processing:</strong> Optimized for vector similarity searches with sub-2ms latency, even for massive datasets.</p></li><li><p><strong>Straightforward API:</strong> Simplifies integration into machine learning workflows, supporting both dense embeddings and sparse embeddings.</p></li></ul><h4><strong>Common Use Cases</strong></h4><ul><li><p><strong>Recommendation Systems:</strong> Leverages semantic similarity to deliver personalized suggestions in e-commerce or content platforms.</p></li><li><p><strong>Retrieval-Augmented Generation (RAG):</strong> Enhances generative AI by retrieving contextually relevant data for large language models.</p></li><li><p><strong>Semantic Search Applications:</strong> Enables complex queries by understanding semantic meaning, ideal for knowledge base retrieval or question answering.</p></li></ul><h4><strong>Strengths</strong></h4><ul><li><p>Exceptional performance metrics for nearest neighbor searches.</p></li><li><p>Supports high throughput workloads with minimal compute resource overhead.</p></li></ul><h4><strong>Limitations</strong></h4><ul><li><p>Proprietary platform limits flexibility compared to open-source alternatives.</p></li><li><p>Deployment options are restricted to cloud environments.</p></li></ul><h2><strong>Weaviate</strong></h2><p>Weaviate is an open-source vector database that combines hybrid search capabilities with multi-modal data support. It bridges the gap between structured data and unstructured data, enabling semantic search across diverse data types like text embeddings and images.</p><h4><strong>Key Features</strong></h4><ul><li><p><strong>Hybrid Search:</strong> Combines vector similarity searches with scalar filtering to handle both structured and unstructured queries effectively.</p></li><li><p><strong>GraphQL API:</strong> Offers a user-friendly query interface for managing metadata and performing complex queries.</p></li><li><p><strong>Multi-modal Support:</strong> Handles diverse data types, including images and text embeddings, making it versatile for AI applications.</p></li></ul><h4><strong>Common Use Cases</strong></h4><ul><li><p><strong>Semantic Search Applications:</strong> Ideal for retrieving semantically similar results in domains like natural language processing or generative AI.</p></li><li><p><strong>Multi-modal Data Applications:</strong> Supports combining text, images, and metadata for tasks like image classification or cross-modal retrieval.</p></li><li><p><strong>Knowledge Base Retrieval:</strong> Enables precise question answering by indexing vast amounts of structured and unstructured data.</p></li></ul><h4><strong>Strengths</strong></h4><ul><li><p>Open-source flexibility allows users to self-host or deploy on the cloud based on specific project requirements.</p></li><li><p>Strong support for hybrid search enables complex queries across metadata-rich datasets.</p></li></ul><h4><strong>Limitations</strong></h4><ul><li><p>Slightly higher latency compared to Pinecone in real-time scenarios.</p></li><li><p>Requires more manual configuration for index creation and performance tuning.</p></li></ul><h2><strong>Milvus</strong></h2><p>Milvus is an open-source vector database purpose-built for large-scale AI applications. Its modular architecture supports diverse index types, making it highly adaptable to various machine learning applications requiring efficient similarity searches.</p><h4><strong>Key Features</strong></h4><ul><li><p><strong>Scalability:</strong> Designed to handle trillions of high-dimensional vectors, making it suitable for massive datasets.</p></li><li><p><strong>Diverse Index Types:</strong> Supports HNSW, IVF, and other indexing methods to optimize nearest neighbor searches based on specific use cases.</p></li><li><p><strong>Flexible Deployment Options:</strong> Offers both self-hosted and cloud-native setups to meet varying infrastructure needs.</p></li></ul><h4><strong>Common Use Cases</strong></h4><ul><li><p><strong>Embeddings Management:</strong> Handles dense embeddings generated by machine learning models for tasks like semantic similarity or recommendation systems.</p></li><li><p><strong>Similarity Search at Scale:</strong> Powers large-scale AI applications like image retrieval or anomaly detection in IoT sensor data.</p></li><li><p><strong>AI Applications in Research:</strong> Supports compute-intensive tasks such as drug discovery or climate modeling through efficient vector search.</p></li></ul><h4><strong>Strengths</strong></h4><ul><li><p>Exceptional scalability makes it a top choice for organizations managing vast amounts of vector data.</p></li><li><p>Modular design allows customization of index creation and query optimization based on performance metrics.</p></li></ul><h4><strong>Limitations</strong></h4><ul><li><p>Steeper learning curve compared to Pinecone and Weaviate due to its modular architecture.</p></li><li><p>Requires expertise in infrastructure management to achieve optimal performance.</p></li></ul><h2><strong>Comparison Summary</strong></h2><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Tc9c!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8bdeb93b-0e0a-414d-b3ba-89def83dcca5_764x556.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Tc9c!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8bdeb93b-0e0a-414d-b3ba-89def83dcca5_764x556.png 424w, https://substackcdn.com/image/fetch/$s_!Tc9c!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8bdeb93b-0e0a-414d-b3ba-89def83dcca5_764x556.png 848w, https://substackcdn.com/image/fetch/$s_!Tc9c!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8bdeb93b-0e0a-414d-b3ba-89def83dcca5_764x556.png 1272w, https://substackcdn.com/image/fetch/$s_!Tc9c!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8bdeb93b-0e0a-414d-b3ba-89def83dcca5_764x556.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Tc9c!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8bdeb93b-0e0a-414d-b3ba-89def83dcca5_764x556.png" width="764" height="556" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8bdeb93b-0e0a-414d-b3ba-89def83dcca5_764x556.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:556,&quot;width&quot;:764,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:38225,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.ankursnewsletter.com/i/158525191?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8bdeb93b-0e0a-414d-b3ba-89def83dcca5_764x556.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Tc9c!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8bdeb93b-0e0a-414d-b3ba-89def83dcca5_764x556.png 424w, https://substackcdn.com/image/fetch/$s_!Tc9c!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8bdeb93b-0e0a-414d-b3ba-89def83dcca5_764x556.png 848w, https://substackcdn.com/image/fetch/$s_!Tc9c!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8bdeb93b-0e0a-414d-b3ba-89def83dcca5_764x556.png 1272w, https://substackcdn.com/image/fetch/$s_!Tc9c!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8bdeb93b-0e0a-414d-b3ba-89def83dcca5_764x556.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Each database offers unique strengths tailored to specific project requirements. <strong>Pinecone is ideal for real-time processing with minimal setup; Weaviate excels in hybrid search across multi-modal data; Milvus is unmatched in scalability for handling massive datasets.</strong> </p><h2><strong>Strengths and Weaknesses</strong></h2><h2><strong>Pinecone</strong></h2><p><strong>Strengths</strong></p><ul><li><p>Fully managed service eliminates operational overhead.</p></li><li><p>Exceptional performance for real-time applications.</p></li><li><p>Simple integration with machine learning workflows.</p></li></ul><p><strong>Weaknesses</strong></p><ul><li><p>Proprietary licensing limits customizability.</p></li><li><p>Higher costs for large-scale deployments compared to open-source options.</p></li></ul><h2><strong>Weaviate</strong></h2><p><strong>Strengths</strong></p><ul><li><p>Open-source flexibility with strong hybrid search capabilities.</p></li><li><p>Built-in support for multi-modal data and customizable models.</p></li><li><p>GraphQL API simplifies querying.</p></li></ul><p><strong>Weaknesses</strong></p><ul><li><p>Limited transaction support (eventual consistency in distributed setups.</p></li><li><p>Slightly higher latency compared to Pinecone and Milvus.</p></li></ul><h2><strong>Milvus</strong></h2><p><strong>Strengths</strong></p><ul><li><p>Highly flexible with diverse indexing options tailored to specific needs.</p></li><li><p>Scalable architecture supports massive datasets (trillion-scale vectors).</p></li><li><p>Strong community support and open-source ecosystem.</p></li></ul><p><strong>Weaknesses</strong></p><ul><li><p>Requires manual configuration for optimal performance.</p></li><li><p>Learning curve can be steeper due to its modular design.</p></li></ul><h2><strong>Performance Benchmarks</strong></h2><h2><strong>Latency and Recall Rates</strong></h2><ul><li><p>Pinecone excels in low-latency scenarios (&lt;2ms) with high recall rates.</p></li><li><p>Milvus offers competitive recall but may require tuning to minimize latency.</p></li><li><p>Weaviate performs well (&lt;100ms) but is slightly slower than Pinecone and Milvus for massive datasets.</p></li></ul><h2><strong>Use Case Suitability</strong></h2><h2><strong>When to Choose Pinecone</strong></h2><ul><li><p>Real-time applications requiring ultra-low latency (e.g., recommendation systems).</p></li><li><p>Teams prioritizing ease of use over customizability.</p></li></ul><h2><strong>When to Choose Weaviate</strong></h2><ul><li><p>Projects needing hybrid search across structured and unstructured data.</p></li><li><p>Teams preferring open-source solutions with multi-modal capabilities.</p></li></ul><h2><strong>When to Choose Milvus</strong></h2><ul><li><p>Large-scale AI applications managing trillions of vectors.</p></li><li><p>Developers needing flexibility in indexing and deployment configurations.</p></li></ul><h2><strong>Cost Considerations</strong></h2><h2><strong>Pinecone</strong></h2><ul><li><p>Premium pricing model; suitable for enterprise-grade projects requiring managed services.</p></li></ul><h2><strong>Weaviate</strong></h2><ul><li><p>The open-source option reduces costs but may require investment in infrastructure if self-hosted.</p></li></ul><h2><strong>Milvus</strong></h2><ul><li><p>Cost-effective for large-scale deployments due to its open-source nature but requires operational expertise.</p></li></ul><div><hr></div><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!-kh9!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!-kh9!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!-kh9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png" width="1456" height="246" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:246,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:172293,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:&quot;&quot;,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!-kh9!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p>I also host an AI podcast and content series called &#8220;Pioneers.&#8221; This series takes you on an enthralling journey into the minds of AI visionaries, founders, and CEOs who are at the forefront of innovation through AI in their organizations.</p><p>To learn more, please visit <a href="https://www.multimodal.dev/blog">Pioneers</a> on Beehiiv.</p><div><hr></div><h2><strong>Wrapping Up</strong></h2><p>Vector databases like Pinecone, Weaviate, and Milvus have become indispensable for AI applications, offering powerful solutions for managing high-dimensional vector data. Each database brings unique strengths to the table, catering to different use cases and project requirements.</p><ul><li><p><strong>Pinecone</strong> stands out as a fully managed, cloud-native vector database designed for real-time processing. Its low-latency performance and minimal setup make it ideal for recommendation systems, semantic search applications, and retrieval-augmented generation (RAG). However, its proprietary nature and cloud-only deployment may limit flexibility for some users.<br></p></li><li><p><strong>Weaviate</strong> shines with its hybrid search capabilities, combining vector similarity searches with structured filtering. Its open-source flexibility and multi-modal data support make it perfect for projects involving both structured and unstructured data, such as knowledge base retrieval or multi-modal applications. While versatile, it may require more manual configuration and has slightly higher latency compared to Pinecone.<br></p></li><li><p><strong>Milvus</strong> is the go-to choice for large-scale AI applications requiring scalability and customization. Its modular architecture supports diverse indexing methods, making it highly adaptable for massive datasets and compute-intensive tasks like embeddings management or similarity search at scale. However, its steep learning curve and need for operational expertise may pose challenges for some teams.</p></li></ul><p><strong>Ultimately, the choice between Pinecone, Weaviate, and Milvus depends on your specific project requirements. Consider factors like deployment preferences (cloud vs. self-hosted), query complexity, scalability needs, and budget constraints.</strong> For real-time performance with minimal overhead, Pinecone is a strong contender. For open-source flexibility and hybrid search capabilities, Weaviate is an excellent option. And for handling massive datasets with advanced indexing options, Milvus is unmatched.</p><p>By carefully evaluating your use case&#8212;whether it's semantic search, recommendation systems, or generative AI&#8212;you can select the right vector database to unlock the full potential of your machine learning models.</p><p>I&#8217;ll come back next week with more insights on building AI for enterprises.</p><p>Until then,</p><p>Ankur.</p>]]></content:encoded></item><item><title><![CDATA[AgentFlow vs Crew AI vs Autogen vs LangChain for Building AI Agents]]></title><description><![CDATA[Explore the top AI agent platforms of 2025: AgentFlow, AutoGen, LangChain, and CrewAI. Compare features, use cases, and implementation considerations for enterprise automation.]]></description><link>https://www.ankursnewsletter.com/p/agentflow-vs-crew-ai-vs-autogen-vs</link><guid isPermaLink="false">https://www.ankursnewsletter.com/p/agentflow-vs-crew-ai-vs-autogen-vs</guid><dc:creator><![CDATA[Ankur A. Patel]]></dc:creator><pubDate>Thu, 27 Feb 2025 16:04:13 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!jdyq!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcc1fdbd1-d84b-4282-814f-c45c6c84aa82_1200x600.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!jdyq!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcc1fdbd1-d84b-4282-814f-c45c6c84aa82_1200x600.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!jdyq!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcc1fdbd1-d84b-4282-814f-c45c6c84aa82_1200x600.png 424w, https://substackcdn.com/image/fetch/$s_!jdyq!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcc1fdbd1-d84b-4282-814f-c45c6c84aa82_1200x600.png 848w, https://substackcdn.com/image/fetch/$s_!jdyq!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcc1fdbd1-d84b-4282-814f-c45c6c84aa82_1200x600.png 1272w, https://substackcdn.com/image/fetch/$s_!jdyq!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcc1fdbd1-d84b-4282-814f-c45c6c84aa82_1200x600.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!jdyq!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcc1fdbd1-d84b-4282-814f-c45c6c84aa82_1200x600.png" width="1200" height="600" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/cc1fdbd1-d84b-4282-814f-c45c6c84aa82_1200x600.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:600,&quot;width&quot;:1200,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:390408,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.ankursnewsletter.com/i/158038267?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcc1fdbd1-d84b-4282-814f-c45c6c84aa82_1200x600.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!jdyq!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcc1fdbd1-d84b-4282-814f-c45c6c84aa82_1200x600.png 424w, https://substackcdn.com/image/fetch/$s_!jdyq!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcc1fdbd1-d84b-4282-814f-c45c6c84aa82_1200x600.png 848w, https://substackcdn.com/image/fetch/$s_!jdyq!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcc1fdbd1-d84b-4282-814f-c45c6c84aa82_1200x600.png 1272w, https://substackcdn.com/image/fetch/$s_!jdyq!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcc1fdbd1-d84b-4282-814f-c45c6c84aa82_1200x600.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.ankursnewsletter.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Subscribe today to join our community of 1880 AI enthusiasts and business leaders for more insider takes and practical advice on implementing AI in business.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h2>Key Takeaways</h2><ul><li><p>AI agent frameworks have evolved from <strong>single-task bots to sophisticated process automation engines</strong> capable of handling complex workflows across multiple systems.</p></li></ul><ul><li><p><a href="https://www.multimodal.dev/agentflow">AgentFlow</a> excels in regulated industries with its <strong>SOC2 Type II compliance</strong>, <strong>built-in audit trails, and industry-specific optimizations for regulated spaces like finance/insurance.</strong></p></li></ul><ul><li><p>AutoGen is best suited for <strong>research teams and experimental AI projects</strong>, offering powerful multi-agent conversation orchestration and LLM optimization capabilities.</p></li></ul><ul><li><p>LangChain provides a <strong>modular ecosystem for developers</strong> to build custom chat interfaces and integrate with various LLM providers, making it ideal for creating tailored AI solutions.</p></li></ul><ul><li><p>CrewAI specializes in human-AI collaboration, particularly for <strong>content creation and research tasks</strong>, with its role-based agent teams and hierarchical process management.</p></li></ul><div><hr></div><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!R9ua!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!R9ua!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!R9ua!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png" width="1456" height="246" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/bea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:246,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:273651,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:&quot;&quot;,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!R9ua!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1456w" sizes="100vw" loading="lazy" fetchpriority="high"></picture><div></div></div></a></figure></div><p>In 2023, I started Multimodal, a Generative AI company that helps organizations automate complex, knowledge-based workflows using AI Agents. Check it out <a href="https://www.multimodal.dev/">here</a>.</p><div><hr></div><p>The evolution of AI agent frameworks has transformed single-task bots into process automation engines capable of orchestrating complex workflows across multiple systems. <strong>Modern multi-agent systems now combine highly specialized agents&#8212;trained on domain-specific training data&#8212;to tackle resource-intensive challenges like supply chain management and customer interactions.</strong></p><p>When building enterprise-grade AI agents in 2025, four criteria dominate:</p><p>1. <strong>Security</strong>: Safeguarding sensitive data while processing enterprise data</p><p>2. <strong>Vertical specialization</strong>: Deploying autonomous agents fine-tuned for specific tasks like inferring customer intent</p><p>3. <strong>Collaboration</strong>: Enabling AI agents to streamline workflows through LLM-powered coordination</p><p>4. <strong>Deployment flexibility</strong>: Integrating generative AI with existing enterprise systems without costly overhauls</p><p>Today's agentic AI platforms leverage large language models to interpret natural language requests, execute complex tasks, and deliver valuable insights through user interfaces. Unlike early AI-powered agents limited to predefined rules, modern frameworks use machine learning to adapt to user preferences, achieving operational efficiency gains of 40-60%.</p><p>Let&#8217;s dive deep into AgentFlow, Crew AI, Langchain, and Autogen to analyze which platform wins today.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!UvLZ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb0729aee-13b3-4b4f-bf13-146114b9709d_1764x527.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!UvLZ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb0729aee-13b3-4b4f-bf13-146114b9709d_1764x527.png 424w, https://substackcdn.com/image/fetch/$s_!UvLZ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb0729aee-13b3-4b4f-bf13-146114b9709d_1764x527.png 848w, https://substackcdn.com/image/fetch/$s_!UvLZ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb0729aee-13b3-4b4f-bf13-146114b9709d_1764x527.png 1272w, https://substackcdn.com/image/fetch/$s_!UvLZ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb0729aee-13b3-4b4f-bf13-146114b9709d_1764x527.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!UvLZ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb0729aee-13b3-4b4f-bf13-146114b9709d_1764x527.png" width="1456" height="435" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b0729aee-13b3-4b4f-bf13-146114b9709d_1764x527.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:435,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:88824,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.ankursnewsletter.com/i/158038267?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb0729aee-13b3-4b4f-bf13-146114b9709d_1764x527.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!UvLZ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb0729aee-13b3-4b4f-bf13-146114b9709d_1764x527.png 424w, https://substackcdn.com/image/fetch/$s_!UvLZ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb0729aee-13b3-4b4f-bf13-146114b9709d_1764x527.png 848w, https://substackcdn.com/image/fetch/$s_!UvLZ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb0729aee-13b3-4b4f-bf13-146114b9709d_1764x527.png 1272w, https://substackcdn.com/image/fetch/$s_!UvLZ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb0729aee-13b3-4b4f-bf13-146114b9709d_1764x527.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h2>Platform Deep Dives</h2><h3>Multimodal AgentFlow</h3><h4>1. Architectural Advantages</h4><p><a href="https://www.multimodal.dev/agentflow">AgentFlow</a> operates as an API-first orchestration layer coordinating four specialized AI agents:</p><p>- <strong>Process Agents</strong>: Automate document classification and data extraction from 100+ submission types</p><p>- <strong>Search Agents</strong>: Cross-reference enterprise databases and third-party APIs</p><p>- <strong>Decide Agents</strong>: Apply rule-based logic for claims approval/denial decisions</p><p>- <strong>Create Agents</strong>: Generate audit-ready reports and compliance documentation</p><p>The platform&#8217;s six modules (Configure &#8594; Orchestrate &#8594; Fine-tune &#8594; Ingest &#8594; Monitor &#8594; Review) enable full workflow automation while maintaining human oversight. Private deployments through AWS/Azure/GCP marketplaces to ensure zero data exfiltration for sensitive financial and insurance workflows.</p><p><strong>Multi-Agent Orchestration</strong>: One of AgentFlow&#8217;s standout features is that it seamlessly orchestrates AI Agents with human supervisors and third party applications.</p><h4>2. Vertical Dominance</h4><p>AgentFlow dominates in the insurance and finance verticals. It&#8217;s purpose built to address multiple workflows within insurance and finance. Here&#8217;s an example of what it can do:<br></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!mDuD!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbc68d01b-e20e-4501-961b-e42704534410_1102x540.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!mDuD!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbc68d01b-e20e-4501-961b-e42704534410_1102x540.png 424w, https://substackcdn.com/image/fetch/$s_!mDuD!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbc68d01b-e20e-4501-961b-e42704534410_1102x540.png 848w, https://substackcdn.com/image/fetch/$s_!mDuD!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbc68d01b-e20e-4501-961b-e42704534410_1102x540.png 1272w, https://substackcdn.com/image/fetch/$s_!mDuD!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbc68d01b-e20e-4501-961b-e42704534410_1102x540.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!mDuD!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbc68d01b-e20e-4501-961b-e42704534410_1102x540.png" width="1102" height="540" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/bc68d01b-e20e-4501-961b-e42704534410_1102x540.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:540,&quot;width&quot;:1102,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:148768,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.ankursnewsletter.com/i/158038267?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbc68d01b-e20e-4501-961b-e42704534410_1102x540.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!mDuD!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbc68d01b-e20e-4501-961b-e42704534410_1102x540.png 424w, https://substackcdn.com/image/fetch/$s_!mDuD!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbc68d01b-e20e-4501-961b-e42704534410_1102x540.png 848w, https://substackcdn.com/image/fetch/$s_!mDuD!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbc68d01b-e20e-4501-961b-e42704534410_1102x540.png 1272w, https://substackcdn.com/image/fetch/$s_!mDuD!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbc68d01b-e20e-4501-961b-e42704534410_1102x540.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><strong>Insurance Claims Automation</strong></p><p>AgentFlow&#8217;s adjudication workflow processes First Notice of Loss (FNOL) claims through:</p><p>1. <strong>Document AI</strong>: Classifies web portal/email submissions into 40+ schemas</p><p>2. <strong>Database AI</strong>: Helps with diligence and adjudication</p><p>3. <strong>Decision AI</strong>: Applies regional compliance rules to approve/deny claims</p><p>4. <strong>Report AI</strong>: Generates adjustment-ready documentation with explainable outputs</p><p><strong>Financial Services</strong></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!ELpC!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fea7e2c52-038d-445c-8f57-7257d412b958_1078x535.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!ELpC!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fea7e2c52-038d-445c-8f57-7257d412b958_1078x535.png 424w, https://substackcdn.com/image/fetch/$s_!ELpC!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fea7e2c52-038d-445c-8f57-7257d412b958_1078x535.png 848w, https://substackcdn.com/image/fetch/$s_!ELpC!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fea7e2c52-038d-445c-8f57-7257d412b958_1078x535.png 1272w, https://substackcdn.com/image/fetch/$s_!ELpC!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fea7e2c52-038d-445c-8f57-7257d412b958_1078x535.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!ELpC!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fea7e2c52-038d-445c-8f57-7257d412b958_1078x535.png" width="1078" height="535" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/ea7e2c52-038d-445c-8f57-7257d412b958_1078x535.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:535,&quot;width&quot;:1078,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:144036,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://www.ankursnewsletter.com/i/158038267?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fea7e2c52-038d-445c-8f57-7257d412b958_1078x535.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!ELpC!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fea7e2c52-038d-445c-8f57-7257d412b958_1078x535.png 424w, https://substackcdn.com/image/fetch/$s_!ELpC!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fea7e2c52-038d-445c-8f57-7257d412b958_1078x535.png 848w, https://substackcdn.com/image/fetch/$s_!ELpC!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fea7e2c52-038d-445c-8f57-7257d412b958_1078x535.png 1272w, https://substackcdn.com/image/fetch/$s_!ELpC!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fea7e2c52-038d-445c-8f57-7257d412b958_1078x535.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The platform automates loan origination through</p><p>- Custom-configured Process Agents extracting data from 20+ document types</p><p>- Decide Agents cross-referencing credit databases with regulatory requirements</p><p>- Create Agents generating SEC-compliant underwriting reports</p><h4>3. Technical Differentiators</h4><p>- <strong>Confidence Scoring</strong>: Flags low-certainty outputs (e.g., ambiguous claim photos) for human review</p><p>- <strong>Self-Learning Core</strong>: Automatically adjusts model parameters based on historical feedback</p><p>- <strong>Multi-Agent Monitoring</strong>: Real-time dashboards track error rates (3% baseline), throughput (1,000 tasks/hour), and system resource utilization</p><h3>AutoGen (Microsoft)</h3><h4>1. Framework Overview</h4><p>- Open-source platform for orchestrating multi-agent AI systems</p><p>- Enables creation and management of specialized AI agents</p><p>- Designed for complex, multi-step task automation</p><h4>2. Key Features</h4><p>- Multi-agent conversation orchestration</p><p>- Facilitates collaboration between multiple AI agents</p><p>- Supports both event-driven and request/response interaction patterns</p><p>- LLM optimization via EcoOptiGen</p><p>- <strong>Reduces LLM usage costs by 40-60%</strong> through intelligent hyperparameter tuning</p><p>- Asynchronous messaging architecture</p><p>- Improves scalability and responsiveness of agent networks</p><h4>3. Technical Capabilities</h4><p>- Modular and extensible design</p><p>- Pluggable components for custom agents, tools, memory, and models</p><p>- Built-in observability and debugging tools</p><p>- <strong>Metric tracking, message tracing, and OpenTelemetry support</strong></p><p>- Cross-language support</p><p>- Interoperability between agents built in different programming languages (Python, .NET)</p><h4>4. Use Cases</h4><p>- Tech R&amp;D and prototyping</p><p>- Ideal for exploring complex topics and building AI research applications</p><p>- Conversational AI development</p><p>- <strong>Enables creation of sophisticated chatbots and dialogue systems</strong></p><p>- Automated project management</p><p>- Streamlines coordination of complex workflows across multiple systems</p><h4>5. Limitations</h4><p>- Steep learning curve for non-developers</p><p>- Requires significant software engineering expertise (300+ lines of code for basic setups)</p><p>- Limited enterprise security features</p><p>- <strong>Lacks built-in protocols for handling sensitive data in production environments</strong></p><p>- High operational costs</p><p>- Significant expenses when scaling to complex tasks, especially with GPT-4 Turbo</p><h4>6. Future Developments</h4><p>- Enhanced low-code interface (AutoGen Studio)</p><p>- Visual workflow builder and real-time agent updates</p><p>- <strong>Improved reasoning capabilities</strong></p><p>- Addressing current limitations in complex multihop question answering</p><p>- Expanded third-party integrations</p><p>- Growing ecosystem of community-developed extensions and tools</p><h3>LangChain</h3><h4>1. Framework Overview</h4><p>LangChain is an open-source framework designed to simplify the development of applications powered by large language models (LLMs). <strong>It provides modular tools for building, deploying, and managing AI systems, enabling developers to create multi-agent systems capable of handling complex workflows.</strong> </p><p>By leveraging LangGraph, LangChain allows for the creation of stateful workflows and agentic AI systems that excel at performing specialized tasks like document analysis, customer interactions, and data retrieval.</p><h4>2. Strengths</h4><p>One of LangChain&#8217;s key strengths is its modular ecosystem, which supports seamless integration with multiple LLM providers such as OpenAI and Hugging Face. This flexibility enables developers to build AI agents tailored to specific tasks by combining LLMs with external tools like APIs or databases. </p><p>LangGraph, a core component of LangChain, facilitates stateful workflows by maintaining memory across interactions, making it ideal for automating tasks that require context retention or multi-step reasoning.</p><p><strong>LangChain also excels in enabling developers to fine-tune agents for specialized tasks using training data and user feedback.</strong> For example, a chatbot built with LangChain can infer customer intent through natural language processing while retrieving relevant enterprise data from knowledge bases or external APIs. This makes it particularly effective for applications requiring advanced retrieval-augmented generation (RAG) techniques or robotic process automation.</p><h4>3. Limitations</h4><p>While LangChain offers powerful tools for building agentic AI systems, it lacks built-in compliance features for handling sensitive data such as patient records or financial information. </p><p><strong>Developers must implement custom security measures to ensure alignment with enterprise requirements.</strong> Additionally, creating full process automation often requires stitching together multiple components, which can be resource-intensive and demand significant software engineering expertise.</p><h4>4. Use Cases</h4><p>LangChain is widely used for developer-centric applications such as chatbots and document analysis systems. For instance, customer service chatbots built with LangChain can handle complex queries by integrating LLMs with external knowledge sources to provide accurate responses. </p><p><strong>In document analysis, LangChain-powered agents process raw data from multiple systems, extract valuable insights, and streamline workflows across entire organizations.</strong> Its flexibility also makes it suitable for coding assistants, marketing automation, and e-commerce personalization.</p><p>By enabling developers to build their own AI agents with modular components and advanced stateful workflows, LangChain continues to be a preferred choice for organizations seeking scalable AI solutions that integrate seamlessly into existing enterprise systems.</p><h3>CrewAI</h3><h4>1. Framework Overview</h4><p>CrewAI is an open-source platform for orchestrating role-playing, autonomous AI agents. It enables developers to build specialized teams of AI agents, each with unique skills, to solve complex tasks through collaboration.</p><h4>2. Key Strengths</h4><p>- <strong>Human-in-the-loop Integration</strong>: CrewAI excels in incorporating human expertise into AI workflows. By setting the `human input` flag, agents can request additional information or clarification from users, enhancing accuracy in complex decision-making processes.</p><p>- <strong>Role-based Agent Teams</strong>: The framework supports creating AI agents with defined roles, goals, and backstories. This approach allows for flexible task delegation and specialized problem-solving.</p><h4>3. Core Features</h4><p>- <strong>Hierarchical Process Management</strong>: CrewAI implements a structured approach to task management, simulating traditional organizational hierarchies for efficient task delegation and execution.</p><p>- <strong>Task Delegation</strong>: A manager agent allocates tasks among crew members based on their roles and capabilities, optimizing workflow efficiency.</p><p>- <strong>Result Validation</strong>: The manager evaluates outcomes to ensure they meet required standards, maintaining quality and accuracy.</p><h4>4. Limitations</h4><p>- <strong>Lack of Multimodal Support:</strong> CrewAI currently doesn't natively support processing both text and images, limiting its application in certain scenarios.</p><p>- <strong>Deployment Constraints</strong>: The framework has limited deployment options, which may restrict its use in some enterprise environments.</p><h4>5. Use Cases</h4><p>- <strong>Content Creation</strong>: CrewAI can automate various aspects of content production, from research to writing and optimization.</p><p>- <strong>Research Collaboratives</strong>: The platform's ability to coordinate multiple specialized agents makes it ideal for complex research tasks.</p><p>- <strong>Customer Service</strong>: CrewAI can handle tasks like call classification, intent discovery, and response suggestions to streamline support operations.</p><h4>6. Technical Considerations</h4><p>CrewAI is built on top of LangChain, offering a modular design that connects smoothly with various tools and APIs. While it provides powerful capabilities for collaborative AI development, it requires some technical expertise to fully utilize. The platform currently lacks a visual builder or extensive no-code options, which may limit accessibility for non-technical users.</p><h2>Implementation Considerations</h2><h3>Choose AgentFlow when:</h3><p>1. <strong>Operating in regulated industries</strong>: AgentFlow's SOC2 Type II compliant architecture ensures secure handling of sensitive data, with built-in audit trails and explainable AI features crucial for regulatory compliance.</p><p>2. <strong>Requiring rapid vendor onboarding</strong>: AgentFlow's API-first platform enables deployment within 90 days, significantly faster than traditional enterprise software implementations.</p><p>3. <strong>Needing real human-AI handoff points</strong>: AgentFlow's orchestration layer seamlessly integrates human supervisors with AI agents, allowing for smooth transitions in complex workflows.</p><h3>Alternative options</h3><p>1. <strong>AutoGen</strong>: Ideal for research teams focused on LLM fine-tuning and multi-agent conversation orchestration. Best suited for experimental AI projects and prototyping.</p><p>2. <strong>LangChain</strong>: Preferred by developer shops building custom chat interfaces. Its modular ecosystem supports flexible integration with various LLM providers.</p><p>3. <strong>CrewAI</strong>: Suited for startups prototyping human-AI collaboration, particularly in content creation and research tasks.</p><p>When implementing agentic AI systems, consider:</p><p>- Complexity of tasks and workflows</p><p>- Security and compliance requirements</p><p>- Integration needs with existing enterprise systems</p><p>- Level of customization required</p><p>- Available technical expertise</p><p>- Scalability and performance demands</p><p>Choosing the right platform depends on balancing these factors against your organization's specific needs and resources.</p><div><hr></div><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!-kh9!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!-kh9!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!-kh9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png" width="1456" height="246" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:246,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:172293,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:&quot;&quot;,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!-kh9!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p>I also host an AI podcast and content series called &#8220;Pioneers.&#8221; This series takes you on an enthralling journey into the minds of AI visionaries, founders, and CEOs who are at the forefront of innovation through AI in their organizations.</p><p>To learn more, please visit <a href="https://www.multimodal.dev/blog">Pioneers</a> on Beehiiv.</p><div><hr></div><h2>Wrapping Up</h2><p>I believe the AI agent landscape in 2025 offers a diverse array of solutions for enterprises seeking to automate complex workflows. While each platform has its strengths, <strong>AgentFlow stands out for its robust security features and industry-specific optimizations, particularly in finance, insurance, and even healthcare.</strong> However, AutoGen, LangChain, and CrewAI each fill important niches for research, development, and collaborative AI projects.</p><p>Ultimately, I believe the success of AI agent implementations will hinge on choosing the right tool for specific organizational needs, balancing factors like compliance requirements, technical expertise, and scalability demands. The potential for transformative efficiency gains makes this an exciting space to watch in the coming years.</p><p>I&#8217;ll come back next week with more on Agentic AI.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.ankursnewsletter.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">If you liked reading this post, consider subscribing for more AI content.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Azure ML vs Vertex AI vs SageMaker: A Comparison]]></title><description><![CDATA[Compare Azure ML, Vertex AI, and SageMaker across key features, use cases, and pricing.]]></description><link>https://www.ankursnewsletter.com/p/azure-ml-vs-vertex-ai-vs-sagemaker</link><guid isPermaLink="false">https://www.ankursnewsletter.com/p/azure-ml-vs-vertex-ai-vs-sagemaker</guid><dc:creator><![CDATA[Ankur A. Patel]]></dc:creator><pubDate>Thu, 13 Feb 2025 17:15:24 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!WQgm!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd2c1b4ad-52ff-4a14-8fcd-fa92063aad61_1200x600.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!WQgm!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd2c1b4ad-52ff-4a14-8fcd-fa92063aad61_1200x600.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!WQgm!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd2c1b4ad-52ff-4a14-8fcd-fa92063aad61_1200x600.png 424w, https://substackcdn.com/image/fetch/$s_!WQgm!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd2c1b4ad-52ff-4a14-8fcd-fa92063aad61_1200x600.png 848w, https://substackcdn.com/image/fetch/$s_!WQgm!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd2c1b4ad-52ff-4a14-8fcd-fa92063aad61_1200x600.png 1272w, https://substackcdn.com/image/fetch/$s_!WQgm!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd2c1b4ad-52ff-4a14-8fcd-fa92063aad61_1200x600.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!WQgm!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd2c1b4ad-52ff-4a14-8fcd-fa92063aad61_1200x600.png" width="1200" height="600" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d2c1b4ad-52ff-4a14-8fcd-fa92063aad61_1200x600.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:600,&quot;width&quot;:1200,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:381205,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!WQgm!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd2c1b4ad-52ff-4a14-8fcd-fa92063aad61_1200x600.png 424w, https://substackcdn.com/image/fetch/$s_!WQgm!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd2c1b4ad-52ff-4a14-8fcd-fa92063aad61_1200x600.png 848w, https://substackcdn.com/image/fetch/$s_!WQgm!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd2c1b4ad-52ff-4a14-8fcd-fa92063aad61_1200x600.png 1272w, https://substackcdn.com/image/fetch/$s_!WQgm!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd2c1b4ad-52ff-4a14-8fcd-fa92063aad61_1200x600.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.ankursnewsletter.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Subscribe today to join our community of 1850 AI enthusiasts and business leaders for more insider takes and practical advice on implementing AI in business.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h2>Key Takeaways</h2><ol><li><p>Azure ML excels in <strong>regulated industries</strong> and hybrid deployments, prioritizing governance and compliance.</p></li><li><p>Vertex AI shines with its cutting-edge AI, multimodal capabilities, and tight <strong>BigQuery integration</strong>.</p></li><li><p>AWS SageMaker <strong>dominates at petascale</strong>, offering unparalleled performance and generative AI support within the AWS ecosystem.</p></li><li><p>Platform choice <strong>depends heavily on existing infrastructure</strong>, with Azure best for Microsoft shops, Vertex for Google Cloud users, and SageMaker for AWS-native organizations.</p></li><li><p>Focus on <strong>specialized silicon (like Inferentia3 and TPU v5p) and automated compliance</strong> will drive future platform differentiation.</p></li></ol><div><hr></div><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!R9ua!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!R9ua!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!R9ua!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png" width="1456" height="246" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/bea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:246,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:273651,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:&quot;&quot;,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!R9ua!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1456w" sizes="100vw" loading="lazy" fetchpriority="high"></picture><div></div></div></a></figure></div><p>Last year, I started Multimodal, a Generative AI company that helps organizations automate complex, knowledge-based workflows using AI Agents. Check it out <a href="https://www.multimodal.dev/">here</a>.</p><div><hr></div><p>The machine learning landscape has transformed radically since 2018, when AWS SageMaker first democratized cloud-based model development. Fast-forward to 2025: Cloud ML platforms now handle trillion-parameter models, quantum-inspired architectures, and ethical AI guardrails&#8212;all while powering mission-critical enterprise workflows.</p><p><strong>This evolution reflects a shift from experimental "skunkworks" projects to industrialized AI pipelines that drive revenue, compliance, and competitive differentiation.</strong></p><p>AWS (34%) maintains its lead through Inferentia3 optimizations, while Azure (29%) dominates regulated industries with confidential computing. GCP (22%) punches above its weight in AI research, leveraging TPU v5p clusters and BigQuery&#8217;s petabyte-scale analytics.</p><p>Let&#8217;s compare each of these players to see which enterprise deployments they&#8217;re the best for.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!u9sp!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6393a612-2e67-43ed-85a6-ea4aed3d1dc2_1600x553.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!u9sp!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6393a612-2e67-43ed-85a6-ea4aed3d1dc2_1600x553.png 424w, https://substackcdn.com/image/fetch/$s_!u9sp!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6393a612-2e67-43ed-85a6-ea4aed3d1dc2_1600x553.png 848w, https://substackcdn.com/image/fetch/$s_!u9sp!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6393a612-2e67-43ed-85a6-ea4aed3d1dc2_1600x553.png 1272w, https://substackcdn.com/image/fetch/$s_!u9sp!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6393a612-2e67-43ed-85a6-ea4aed3d1dc2_1600x553.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!u9sp!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6393a612-2e67-43ed-85a6-ea4aed3d1dc2_1600x553.png" width="1456" height="503" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/6393a612-2e67-43ed-85a6-ea4aed3d1dc2_1600x553.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:503,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!u9sp!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6393a612-2e67-43ed-85a6-ea4aed3d1dc2_1600x553.png 424w, https://substackcdn.com/image/fetch/$s_!u9sp!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6393a612-2e67-43ed-85a6-ea4aed3d1dc2_1600x553.png 848w, https://substackcdn.com/image/fetch/$s_!u9sp!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6393a612-2e67-43ed-85a6-ea4aed3d1dc2_1600x553.png 1272w, https://substackcdn.com/image/fetch/$s_!u9sp!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6393a612-2e67-43ed-85a6-ea4aed3d1dc2_1600x553.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h2>Platform Architecture Deep Dive</h2><h3>Azure Machine Learning</h3><p>Azure&#8217;s 2025 architecture now ships <a href="https://azure.microsoft.com/en-us/products/ai-foundry">Unified AI Studio</a>, merging generative AI workflows with traditional predictive modeling &#8211; a game-changer for enterprises running mixed AI pipelines. The platform&#8217;s Confidential ML leverages Intel SGX v4 to encrypt model weights during training, achieving 98% accuracy on encrypted health data trials.</p><p><strong>AutoML for time series gets a 2025 boost with multi-horizon forecasting, slashing energy grid prediction errors by 32% using BigQuery-integrated pipelines.</strong> Hybrid deployments shine through Arc-enabled clusters, allowing manufacturers to train global models centrally while inferencing locally at 150+ edge sites.</p><p>The technical stack flexes enterprise-grade muscle:</p><p>- <strong>Compute</strong>: ND H100 vGPUs deliver 1.4 petaflops per cluster, train 70B-parameter models in &lt;8 hours.</p><p>- <strong>Data</strong>: Synapse Analytics integration processes 14M events/sec for real-time retail demand signals.</p><p>- <strong>Security</strong>: Purview-powered lineage tracking now auto-generates GDPR 2025 compliance reports, tracing 100% of model/data interactions.</p><p>This architecture positions Azure ML as the Swiss Army knife for enterprises needing to balance AI innovation with ironclad governance &#8211; particularly in regulated sectors like finance and healthcare.</p><h3>Google Vertex AI</h3><p>Vertex AI cements its position as 2025's most versatile enterprise ML platform through Gemini 2.0's native multimodality, processing text, code, and 4K video frames in unified workflows. The model's 128K-token context window now handles technical manuals and live sensor feeds simultaneously, powering real-time quality control systems for manufacturers like Siemens.</p><p><strong>BigQuery ML direct deployment slashes time-to-production: Data teams build models in SQL, then deploy them as REST endpoints in Vertex AI with one-click registry</strong> &#8211; cutting pharmaceutical trial analysis cycles from weeks to 72 hours. The new Agent Builder toolkit ships pre-built RAG templates that reduced Bloomberg's financial report parsing errors by 41% using:</p><p>- Dynamic document chunking</p><p>- Cross-source fact verification</p><p>- Gemini-powered summary distillation</p><p>Underpinning these features, TPU v5p clusters deliver 2.8X faster training than 2024's v4 pods, while A3 VMs with NVIDIA L40S GPUs slash inference costs 37% via:</p><p>- 4-bit quantization for 70B-parameter models</p><p>- 1.2M tokens/sec throughput</p><p>- 8ms p99 latency at 10K QPS</p><p>The technical stack shines in hybrid environments:</p><p>- <strong>Data</strong>: BigQuery Omni analyzes AWS/Azure data without migration</p><p>- <strong>Compute</strong>: Autopilot scales TPU v5p slices during peak genomics workloads</p><p>- <strong>MLOps</strong>: Vertex Pipelines auto-generates Looker dashboards tracking model drift &amp; GDPR compliance</p><p>This architecture makes Vertex AI the go-to for enterprises balancing cutting-edge AI with operational pragmatism &#8211; particularly those leveraging multi-cloud strategies or regulated data.</p><h3>AWS SageMaker</h3><p>AWS cements its enterprise AI dominance, merging data lakes, ML workflows, and generative AI into a single orchestration layer. The 2025 flagship HyperPod now trains 100B+ parameter models like Amazon Nova with:</p><p>- Automatic fault recovery (99.9% uptime during 6-week training cycles)</p><p>- 2.1 exaflop throughput via EC2 DL2q instances</p><p>- Flexible training plans optimizing $2.3M/month GPU budgets</p><p><strong>Neptune-enhanced graph ML delivers real-time fraud detection at Visa-scale, processing 14M transactions/sec with 93% accuracy</strong>. Meanwhile, Inferentia3 chips slash LLM inference costs 58% using:</p><p>- 8-bit floating point quantization</p><p>- 1.4M tokens/sec throughput</p><p>- Sub-10ms p99 latency</p><p>The technical stack redefines cloud-scale AI:</p><p>- <strong>Data</strong>: Lakehouse Federation queries S3/Redshift without ETL - Pfizer reduced data prep time from weeks to hours</p><p>- <strong>Compute</strong>: DL2q clusters with Qualcomm AI 100 Ultra chips achieve 3.2 petaops/Watt</p><p>- <strong>Security</strong>: IAM Roles Anywhere extends permissions to hybrid edge deployments, enforcing Zero Trust via automated credential rotation</p><p>This architecture positions SageMaker as the brute-force solution for enterprises pushing AI at petascale - particularly those needing to industrialize generative AI while maintaining legacy AWS investments.</p><h2>Enterprise Applications &amp; Case Studies</h2><p>The real-world impact of these platforms becomes clear when examining flagship implementations across industries.</p><h3>Financial Services</h3><p>- <strong>Azure ML</strong>: <a href="https://www.linkedin.com/pulse/ai-driven-solutions-hidden-heroes-bottlenecks-risks-thierry-motsch-oa2ne">BNP Paribas</a> achieved 98.7% fraud detection accuracy using Azure&#8217;s confidential computing and federated learning, processing 14M transactions/sec while keeping sensitive financial data encrypted in-place. Their Purview-powered lineage tracking now auto-generates EU Digital Finance Package compliance reports in &lt;8 seconds.</p><p>- <strong>Vertex AI</strong>: <a href="https://cloud.google.com/customers/hsbc-risk-advisory-tool">HSBC</a>&#8217;s risk advisory tool slashed scenario modeling time by 40% using BigQuery ML direct deployment and TPU v5p clusters. Traders now simulate 16,000 default-risk scenarios simultaneously, optimizing $9B+ fixed-income portfolios in real-time.</p><p>- <strong>SageMaker</strong>: <a href="https://reinvent.awsevents.com/content/dam/reinvent/2024/slides/fsi/FSI315_JPMorganChase-Real-time-fraud-screening-at-massive-scale.pdf">JPMorgan Chase</a> processes $2.4T daily transactions through Inferentia3-powered models, achieving 9ms p99 latency for real-time fraud screening. The system auto-scales to 1.2M QPS during market volatility, reducing false positives by 32%.</p><h3>Healthcare</h3><p>- <strong>Azure ML</strong>: <a href="https://www.biopharmatrend.com/post/1125-mayo-clinic-partners-with-microsoft-and-cerebras-on-radiology-and-genomics-foundation-models/">Mayo Clinic&#8217;s </a>radiology foundation model analyzes 20M+ X-ray images, generating AI-powered reports in 11 seconds with 94.3% diagnostic concordance. The system flags critical findings 23 minutes faster than manual reviews.</p><p>- <strong>Vertex AI</strong>: <a href="https://www.prnewswire.com/news-releases/google-cloud-launches-ai-powered-solutions-to-safely-accelerate-drug-discovery-and-precision-medicine-301825674.html">Pfizer</a> reduced drug discovery cycles from 18 months to 72 days using Vertex&#8217;s Target &amp; Lead ID Suite. Their COVID-19 antiviral candidate screened 1.4M protease inhibitors via BigQuery Omni cross-cloud analytics, achieving 89% target binding affinity.</p><p>- <strong>SageMaker</strong>: <a href="https://www.unitedhealthgroup.com/newsroom/2021/2021-7-8-uhc-predictive-analytics-social-determinants-health.html">UnitedHealth&#8217;s</a> readmission predictor combines 120+ clinical variables with wearable sensor data, achieving 83% precision using SageMaker Feature Store. The model alerts care teams 48hrs pre-discharge, reducing 30-day readmissions by 19%.</p><h2>Deployment Considerations</h2><h3>Scalability Patterns</h3><p>Modern ML platforms handle 1K-1M QPS through adaptive scaling:</p><p>- <strong>Auto-scaling</strong>: AWS SageMaker HyperPod scales to 15,000 nodes for 100B+ parameter models, while Vertex AI&#8217;s TPU v5p clusters achieve 2.8X faster training than 2024 standards.</p><p>- <strong>Spot strategies</strong>: AWS&#8217; Price-Capacity-Optimized allocation reduces interruptions by 67% while cutting costs 90% vs on-demand. Azure ML&#8217;s Arc-enabled edge deployments support 150+ hybrid sites with &lt;10ms latency.</p><p>- <strong>Cold start fixes</strong>: Pre-warmed containers (Kubernetes daemonsets) and model quantization slash LLM cold starts from 6 minutes to 40 seconds.</p><h3>Security</h3><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!_zNU!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F503a8666-1e8f-4fa0-8259-e868427c4414_1054x138.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!_zNU!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F503a8666-1e8f-4fa0-8259-e868427c4414_1054x138.png 424w, https://substackcdn.com/image/fetch/$s_!_zNU!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F503a8666-1e8f-4fa0-8259-e868427c4414_1054x138.png 848w, https://substackcdn.com/image/fetch/$s_!_zNU!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F503a8666-1e8f-4fa0-8259-e868427c4414_1054x138.png 1272w, https://substackcdn.com/image/fetch/$s_!_zNU!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F503a8666-1e8f-4fa0-8259-e868427c4414_1054x138.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!_zNU!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F503a8666-1e8f-4fa0-8259-e868427c4414_1054x138.png" width="1054" height="138" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/503a8666-1e8f-4fa0-8259-e868427c4414_1054x138.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:138,&quot;width&quot;:1054,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!_zNU!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F503a8666-1e8f-4fa0-8259-e868427c4414_1054x138.png 424w, https://substackcdn.com/image/fetch/$s_!_zNU!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F503a8666-1e8f-4fa0-8259-e868427c4414_1054x138.png 848w, https://substackcdn.com/image/fetch/$s_!_zNU!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F503a8666-1e8f-4fa0-8259-e868427c4414_1054x138.png 1272w, https://substackcdn.com/image/fetch/$s_!_zNU!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F503a8666-1e8f-4fa0-8259-e868427c4414_1054x138.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><h3>Cost Optimization</h3><p>- <strong>Reserved Instances</strong>: AWS 3-year commitments cut inference costs 58% via Inferentia3 chips, while Azure&#8217;s 1-year reservations save 42% vs on-demand.</p><p>- <strong>Batch vs real-time</strong>: Vertex AI batch predictions cost $0.02/1K data points (&gt;50M scale) vs real-time&#8217;s $2.19M output tokens.</p><p>- <strong>Waste detection</strong>: Tools like CloudPilot AI identify 32% underutilized GPUs via ML-driven resource matching.</p><h2>Platform Selection Guide</h2><h3>Choose Azure ML When</h3><p>- Microsoft ecosystem integration dominates your stack (Teams, Power BI, Dynamics 365)</p><p>- Regulated industries (healthcare/finance) demand 93+ compliance certifications like FedRAMP High and HITRUST</p><p>- Hybrid deployments require Arc-enabled edge clusters &#8211; Siemens uses 150+ sites for real-time quality control with 10ms latency</p><h4>Choose Vertex AI When</h4><p>- <strong>BigQuery is your data backbone</strong></p><p>- Multimodal workflows need Gemini 2.0&#8217;s 128K-token context</p><p>- <strong>Cost-sensitive GenAI</strong> at $0.14/M input tokens vs OpenAI&#8217;s $15/M</p><h4>Choose SageMaker When</h4><p>- Training 100B+ models</p><p>- AWS-native environments</p><p>- Custom silicon optimization</p><p>For multi-cloud strategies, Vertex AI&#8217;s BigQuery Omni analyzes cross-cloud data without migration, while Azure ML&#8217;s Confidential AI protects sensitive workloads across hybrid environments. SageMaker dominates pure scale &#8211; its HyperPod trains models 40% faster than 2024 benchmarks.</p><div><hr></div><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!-kh9!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!-kh9!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!-kh9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png" width="1456" height="246" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:246,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:172293,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:&quot;&quot;,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!-kh9!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p>I also host an AI podcast and content series called &#8220;Pioneers.&#8221; This series takes you on an enthralling journey into the minds of AI visionaries, founders, and CEOs who are at the forefront of innovation through AI in their organizations.</p><p>To learn more, please visit <a href="https://www.multimodal.dev/blog">Pioneers</a> on Beehiiv.</p><div><hr></div><h2>Wrapping Up</h2><p>Enterprises must align platform selection with core infrastructure DNA: <strong>Microsoft shops gain immediate ROI from Azure ML&#8217;s Purview integration, while AWS-native firms exploit SageMaker&#8217;s Inferentia3 chips for $2.4T/day transaction scaling</strong>. Migration demands technical debt audits as well.</p><p>As hybrid architectures become standard, platforms competing on specialized silicon (Inferentia3 vs TPU v5p) and compliance automation will dominate. The next frontier? Quantum-resistant encryption for AI models &#8211; already in Azure ML&#8217;s 2026 roadmap.</p><p>I&#8217;ll come back next week with more such comparisons. <br><br>Until then,<br>Ankur</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.ankursnewsletter.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">If you liked reading this post, consider subscribing for more AI content.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[DeepSeek R1 Vs Open AI O1: Who's the Winner?]]></title><description><![CDATA[DeepSeek R1 challenges OpenAI's o1 with comparable performance at lower costs, marking a potential shift in AI development and accessibility.]]></description><link>https://www.ankursnewsletter.com/p/deepseek-r1-vs-open-ai-o1-whos-the</link><guid isPermaLink="false">https://www.ankursnewsletter.com/p/deepseek-r1-vs-open-ai-o1-whos-the</guid><dc:creator><![CDATA[Ankur A. Patel]]></dc:creator><pubDate>Thu, 30 Jan 2025 16:18:22 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!qACh!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F15d7023e-bf23-4110-b1f0-c334b10024a7_1200x600.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!qACh!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F15d7023e-bf23-4110-b1f0-c334b10024a7_1200x600.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!qACh!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F15d7023e-bf23-4110-b1f0-c334b10024a7_1200x600.png 424w, https://substackcdn.com/image/fetch/$s_!qACh!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F15d7023e-bf23-4110-b1f0-c334b10024a7_1200x600.png 848w, https://substackcdn.com/image/fetch/$s_!qACh!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F15d7023e-bf23-4110-b1f0-c334b10024a7_1200x600.png 1272w, https://substackcdn.com/image/fetch/$s_!qACh!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F15d7023e-bf23-4110-b1f0-c334b10024a7_1200x600.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!qACh!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F15d7023e-bf23-4110-b1f0-c334b10024a7_1200x600.png" width="1200" height="600" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/15d7023e-bf23-4110-b1f0-c334b10024a7_1200x600.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:600,&quot;width&quot;:1200,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:377643,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!qACh!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F15d7023e-bf23-4110-b1f0-c334b10024a7_1200x600.png 424w, https://substackcdn.com/image/fetch/$s_!qACh!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F15d7023e-bf23-4110-b1f0-c334b10024a7_1200x600.png 848w, https://substackcdn.com/image/fetch/$s_!qACh!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F15d7023e-bf23-4110-b1f0-c334b10024a7_1200x600.png 1272w, https://substackcdn.com/image/fetch/$s_!qACh!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F15d7023e-bf23-4110-b1f0-c334b10024a7_1200x600.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.ankursnewsletter.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Subscribe today to join our community of 1840 AI enthusiasts and business leaders for more insider takes and practical advice on implementing AI in business.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h2>Key Takeaways</h2><ol><li><p>DeepSeek R1, an open-source AI model from China, has <strong>matched or surpassed </strong>OpenAI's o1 in various benchmarks while using fewer resources.</p></li><li><p>Both models feature a <strong>128K token context window and multimodal capabilities</strong>, enabling complex reasoning and analysis.</p></li><li><p>DeepSeek R1 excels in <strong>coding tasks</strong>, while OpenAI o1 slightly outperforms in high-level mathematical challenges.</p></li><li><p>DeepSeek R1's <strong>open-source nature</strong> and lower costs make it more accessible, while OpenAI o1 offers robust built-in safety features for enterprise use.</p></li><li><p>These advanced AI models are revolutionizing <strong>scientific research, business innovation, and global technological competition</strong>.</p></li></ol><div><hr></div><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!R9ua!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!R9ua!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!R9ua!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png" width="1456" height="246" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/bea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:246,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:273651,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:&quot;&quot;,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!R9ua!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1456w" sizes="100vw" loading="lazy" fetchpriority="high"></picture><div></div></div></a></figure></div><p>Last year, I started Multimodal, a Generative AI company that helps organizations automate complex, knowledge-based workflows using AI Agents. Check it out <a href="https://www.multimodal.dev/">here</a>.</p><div><hr></div><blockquote><p>&#8220;<a href="https://www.nytimes.com/2025/01/28/technology/china-deepseek-ai-silicon-valley.html">Why DeepSeek Could Change What Silicon Valley Believes About A.I.</a>&#8221;</p><p>"<a href="https://www.npr.org/2025/01/28/g-s1-45061/deepseek-did-a-little-known-chinese-startup-cause-a-sputnik-moment-for-ai">AI's Sputnik Moment</a>"</p></blockquote><p>Headlines like these flooded all our weekends, and in a black-swan event that's sent shockwaves through Silicon Valley and beyond, Chinese startup DeepSeek has unveiled its latest artificial intelligence model, <strong>DeepSeek R1</strong>. This open-source "reasoning" model has not only matched but in some cases surpassed the capabilities of industry leader OpenAI's o1, all while operating on a fraction of the budget and computational resources.</p><p>As we delve deeper into the capabilities, technical specifications, and implications of DeepSeek R1 and OpenAI o1, we'll explore how these models are not just technological marvels, but potential game-changers in the fields of scientific research, business innovation, and global technological competition.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!r6Uw!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77b95c55-b544-4554-a8c4-de584a5df7e8_1798x608.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!r6Uw!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77b95c55-b544-4554-a8c4-de584a5df7e8_1798x608.png 424w, https://substackcdn.com/image/fetch/$s_!r6Uw!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77b95c55-b544-4554-a8c4-de584a5df7e8_1798x608.png 848w, https://substackcdn.com/image/fetch/$s_!r6Uw!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77b95c55-b544-4554-a8c4-de584a5df7e8_1798x608.png 1272w, https://substackcdn.com/image/fetch/$s_!r6Uw!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77b95c55-b544-4554-a8c4-de584a5df7e8_1798x608.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!r6Uw!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77b95c55-b544-4554-a8c4-de584a5df7e8_1798x608.png" width="1456" height="492" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/77b95c55-b544-4554-a8c4-de584a5df7e8_1798x608.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:492,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:162494,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!r6Uw!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77b95c55-b544-4554-a8c4-de584a5df7e8_1798x608.png 424w, https://substackcdn.com/image/fetch/$s_!r6Uw!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77b95c55-b544-4554-a8c4-de584a5df7e8_1798x608.png 848w, https://substackcdn.com/image/fetch/$s_!r6Uw!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77b95c55-b544-4554-a8c4-de584a5df7e8_1798x608.png 1272w, https://substackcdn.com/image/fetch/$s_!r6Uw!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77b95c55-b544-4554-a8c4-de584a5df7e8_1798x608.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h2>Technical specifications</h2><h3>Model architecture</h3><h4>DeepSeek R1</h4><p>- Mixture of Experts (MoE) framework with 671 billion parameters</p><p>- Activates only 37 billion parameters per forward pass</p><p>- Enables specialized processing across various domains</p><p>- Efficient for reasoning</p><h4>OpenAI o1</h4><p>- Advanced architecture incorporating reinforcement learning capabilities</p><p>- Designed for high performance across reasoning benchmarks</p><p>- Specific architectural details limited</p><h3>Training methodology</h3><h4>DeepSeek R1</h4><p>- Combines reinforcement learning and supervised fine-tuning</p><p>- Multi-step process including:</p><p>- Self-verification</p><p>- Accuracy rewards</p><p>- Enhances accuracy and contextual appropriateness of responses</p><h4>OpenAI o1</h4><p>- Employs Reinforcement Learning with Human Feedback (RLHF)</p><p>- Learns from human preferences</p><p>- Aims for more natural and aligned responses in complex reasoning</p><h3>Context window and multimodal capabilities</h3><p>- Both models feature 128K token context window</p><p>- Enables processing of extensive information in single interactions</p><p>- Beneficial for:</p><p>- Long-chain reasoning</p><p>- <strong>Analysis of lengthy documents</strong></p><p>- Multimodal capabilities:</p><p>- Text and image processing</p><p>- Applications in computer vision and natural language understanding</p><h3>Performance benchmarks</h3><h4>Mathematics</h4><h5>DeepSeek R1</h5><p>- 79.8% score on AIME 2024 benchmark</p><p>- Demonstrates strong multi-step reasoning abilities</p><h5>OpenAI o1</h5><p>- 83% score on International Mathematics Olympiad (IMO) qualifying exam</p><p>- Slightly outperforms DeepSeek R1 in high-level mathematical challenges</p><h4>Coding</h4><h5>DeepSeek R1</h5><p>- 96.3 percentile on Codeforces</p><p>- Excels in understanding and generating complex code</p><h5>OpenAI o1</h5><p>- 89th percentile on Codeforces</p><p>- Strong performance, though slightly behind DeepSeek R1</p><h4>General knowledge</h4><h5>DeepSeek R1</h5><p>- 71.5% Pass@1 rate on GPQA Diamond benchmark</p><p>- Demonstrates broad knowledge across diverse topics</p><h5>OpenAI o1</h5><p>- PhD-level performance on physics, chemistry, and biology benchmarks</p><p>- Shows deep understanding of scientific concepts and complex problem-solving</p><h4>Advanced features</h4><h5>DeepSeek R1</h5><p>- Utilizes multi-head latent attention</p><p>- Supports various quantization techniques (8-bit, 4-bit)</p><p>- Open-source nature allows for:</p><p>- <strong>Community-driven development</strong></p><p>- Customization potential</p><p>- Fine-tuning for specific applications</p><h4>OpenAI o1</h4><p>- Specific features less known</p><p>- Strong benchmark performance suggests:</p><p>- Advanced reasoning techniques</p><p>- Novel approaches to natural language processing</p><h2>Applications and use cases</h2><p>The models achieve advanced reasoning capabilities, opening up a wide range of applications across various domains. These large language models, with their complex reasoning tasks and high benchmark performance, are revolutionizing how we approach problem-solving and knowledge-based work.</p><h3>DeepSeek R1</h3><p>DeepSeek R1, an open-source model developed by a Chinese company, has demonstrated strong performance in several key areas:</p><h4>Scientific research and data analysis</h4><p>- Leverages multi-head latent attention for <strong>in-depth scientific reasoning</strong></p><p>- Excels in analyzing complex datasets and drawing insights</p><p>- Capable of performing <strong>long chains of thought</strong> for research hypotheses</p><h4>Advanced code generation and debugging</h4><p>- Achieves high performance on coding benchmarks</p><p>- Utilizes reinforcement learning and supervised fine-tuning for accurate code generation</p><p>- Supports various programming languages and frameworks</p><h4>Complex mathematical problem-solving</h4><p>- Demonstrates advanced reasoning in solving math problems</p><p>- Performs well on math tasks and reasoning benchmarks</p><p>- Capable of <strong>breaking down complex equations</strong> and providing step-by-step solutions</p><h4>Content creation and copywriting</h4><p>- Generates creative writing pieces with coherent narratives</p><p>- Adapts to various writing styles and formats</p><p>- Implements <strong>self-verification</strong> to ensure high-quality outputs</p><h4>Language translation and learning</h4><p>- Supports multiple languages for translation tasks</p><p>- Assists <strong>in language learning</strong> by providing context and explanations</p><p>- Utilizes its large knowledge base for cultural nuances in translations</p><p>DeepSeek's open-source nature allows for community-driven development and customization, making it a versatile tool for researchers and developers. Its API and chat interface provide easy access to its capabilities, enabling integration into various applications.</p><h3>OpenAI o1</h3><p>OpenAI o1, while not open-source, offers a range of advanced features and has been trained exclusively on complex reasoning tasks:</p><h4>Strategy ideation and complex problem-solving</h4><p>- Excels in generating innovative strategies for business and organizational challenges</p><p>- Utilizes advanced reasoning to break down complex problems into manageable steps</p><p>- Capable of considering multiple perspectives and potential outcomes</p><h4>Educational content development and tutoring</h4><p>- Creates comprehensive educational materials across various subjects</p><p>- Adapts explanations to different learning levels</p><p>- Provides interactive tutoring experiences through its chat interface</p><h4>Advanced coding exercises and reviews</h4><p>- Generates challenging coding exercises for skill development</p><p>- Performs code reviews with detailed feedback and suggestions for improvement</p><p>- Assists in optimizing algorithms and improving code efficiency</p><h4>UX design-to-code conversion</h4><p>- Translates UX design concepts into functional code</p><p>- Understands design principles and implements them in various programming languages</p><p>- Provides suggestions for improving user interface and experience</p><h4>Complex writing tasks</h4><p>- Handles sophisticated writing assignments across different genres and styles</p><p>- Implements advanced language models for nuanced and context-aware writing</p><p>- Capable of long-form content creation with coherent structure and argumentation</p><p>Both DeepSeek R1 and OpenAI o1 have undergone rigorous training processes, including reinforcement learning and accuracy rewards, to achieve their advanced reasoning capabilities. <strong>They excel not only in reasoning tasks but also in non-reasoning tasks, demonstrating their versatility.</strong></p><p>The development of these models represents a significant leap forward in AI technology. DeepSeek R1's open-source nature allows for greater transparency and collaborative improvement, while <strong>OpenAI o1's proprietary model offers highly refined and specialized capabilities</strong>.</p><p>As these models continue to evolve, we can expect to see even more sophisticated applications. Future developments may include enhanced cross-checking abilities, improved performance on GPQA diamond and other benchmarks, and more efficient use of reasoning tokens during the forward pass.</p><h2>Cost and accessibility</h2><h3>DeepSeek R1</h3><p>- <strong>Open-source model</strong>: DeepSeek R1 is released under the MIT license, making it freely available for use, modification, and commercial applications.</p><p>- <strong>Cost-effective</strong>: DeepSeek R1's API pricing is significantly lower, charging $0.14 per million input tokens (cache hits) and $2.19 per million output tokens.</p><p>- <strong>Community development</strong>: The open-source nature allows for community-driven improvements and customization.</p><h3>OpenAI o1</h3><p>- <strong>Proprietary model</strong>: o1 is a closed-source model with limited access.</p><p>- <strong>Higher costs</strong>: API pricing for o1 is $15.00 per million input tokens, $7.50 per million cached input tokens, and $60.00 per million output tokens.</p><p>- <strong>Subscription options</strong>: OpenAI offers a "Pro" plan at $200 per month for unlimited access to o1.</p><h2>Deployment options</h2><h3>DeepSeek R1</h3><p>- <strong>Local deployment</strong>: Supports local deployment on various hardware configurations.</p><p>- <strong>Quantization support</strong>: Offers 8-bit and 4-bit quantization for efficient deployment.</p><h3>OpenAI o1</h3><p>- <strong>Cloud-based</strong>: Primarily accessible through API, requiring cloud-based infrastructure.</p><p>- <strong>Limited flexibility</strong>: Does not support on-premise deployment for most users.</p><h2>Ethical considerations and safety features</h2><h3>DeepSeek R1</h3><p>- <strong>Open-source transparency</strong>: Allows for community scrutiny of the model's architecture and training process.</p><p>- <strong>Potential risks</strong>: As an open-source model, it may be more susceptible to misuse if proper safeguards are not implemented.</p><h3>OpenAI o1</h3><p>- <strong>Built-in safety features</strong>: Includes content filtering and bias mitigation mechanisms.</p><p>- <strong>Controlled access</strong>: OpenAI can monitor and regulate usage due to its proprietary nature.</p><div><hr></div><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!-kh9!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!-kh9!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!-kh9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png" width="1456" height="246" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:246,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:172293,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:&quot;&quot;,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!-kh9!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p>I also host an AI podcast and content series called &#8220;Pioneers.&#8221; This series takes you on an enthralling journey into the minds of AI visionaries, founders, and CEOs who are at the forefront of innovation through AI in their organizations.</p><p>To learn more, please visit <a href="https://www.multimodal.dev/blog">Pioneers</a> on Beehiiv.</p><div><hr></div><h2>Wrapping up</h2><p>DeepSeek R1 represents a significant step towards democratizing AI access with its open-source approach and cost-effectiveness. OpenAI o1, while more expensive, offers robust built-in safety features and is tailored for enterprise-level applications.</p><p>It&#8217;ll be interesting to witness how DeepSeek challenges Silicon Valley&#8217;s dominance in AI. I&#8217;ll keep you updated here with more on DeepSeek and the latest in the industry.</p><p>See you next week. <br><br>Until then,</p><p>Ankur.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.ankursnewsletter.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">If you liked reading this post, consider subscribing for more AI content.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Lambda vs RunPod vs Together AI for AI Inference]]></title><description><![CDATA[Dive into a comprehensive comparison of Lambda, RunPod, and Together AI for AI inference. Discover unique strengths, performance metrics, and applications.]]></description><link>https://www.ankursnewsletter.com/p/lambda-vs-runpod-vs-together-ai-for</link><guid isPermaLink="false">https://www.ankursnewsletter.com/p/lambda-vs-runpod-vs-together-ai-for</guid><dc:creator><![CDATA[Ankur A. Patel]]></dc:creator><pubDate>Thu, 23 Jan 2025 14:31:37 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!K5kC!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1780ccc1-e957-4362-9349-957a5f89d8aa_1200x600.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!K5kC!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1780ccc1-e957-4362-9349-957a5f89d8aa_1200x600.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!K5kC!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1780ccc1-e957-4362-9349-957a5f89d8aa_1200x600.png 424w, https://substackcdn.com/image/fetch/$s_!K5kC!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1780ccc1-e957-4362-9349-957a5f89d8aa_1200x600.png 848w, https://substackcdn.com/image/fetch/$s_!K5kC!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1780ccc1-e957-4362-9349-957a5f89d8aa_1200x600.png 1272w, https://substackcdn.com/image/fetch/$s_!K5kC!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1780ccc1-e957-4362-9349-957a5f89d8aa_1200x600.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!K5kC!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1780ccc1-e957-4362-9349-957a5f89d8aa_1200x600.png" width="1200" height="600" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/1780ccc1-e957-4362-9349-957a5f89d8aa_1200x600.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:600,&quot;width&quot;:1200,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:377533,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!K5kC!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1780ccc1-e957-4362-9349-957a5f89d8aa_1200x600.png 424w, https://substackcdn.com/image/fetch/$s_!K5kC!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1780ccc1-e957-4362-9349-957a5f89d8aa_1200x600.png 848w, https://substackcdn.com/image/fetch/$s_!K5kC!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1780ccc1-e957-4362-9349-957a5f89d8aa_1200x600.png 1272w, https://substackcdn.com/image/fetch/$s_!K5kC!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1780ccc1-e957-4362-9349-957a5f89d8aa_1200x600.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.ankursnewsletter.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Subscribe today to join our community of 1830 AI enthusiasts and business leaders for more insider takes and practical advice on implementing AI in business.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h2>Key Takeaways</h2><p>1. Lambda focuses on <strong>high-performance hardware</strong>, RunPod on flexibility and cost-effectiveness, and Together AI on specialized AI offerings and optimizations.</p><p>2. Together AI's Inference Engine claims up to <strong>75% faster performance</strong> than base PyTorch, while RunPod offers sub-250ms cold start times across 30 global regions.</p><p>3. All platforms excel in NLP, <strong>computer vision, audio processing, and multimodal AI</strong>, with Together AI offering 200+ open-source models and RunPod popular for AI art generation.</p><p>4. Together AI provides OpenAI-compatible APIs, RunPod offers a user-friendly CLI, and all platforms support <strong>multiple programming languages and model customization</strong>.</p><p>5. Pricing varies: RunPod uses per-hour GPU instance pricing, Together AI employs a per-token model, and all platforms offer <strong>cost optimization strategies</strong> including serverless options and batch processing.</p><div><hr></div><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!R9ua!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!R9ua!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!R9ua!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png" width="1456" height="246" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/bea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:246,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:273651,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:&quot;&quot;,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!R9ua!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1456w" sizes="100vw" loading="lazy" fetchpriority="high"></picture><div></div></div></a></figure></div><p>Last year, I started Multimodal, a Generative AI company that helps organizations automate complex, knowledge-based workflows using AI Agents. Check it out <a href="https://www.multimodal.dev/">here</a>.</p><div><hr></div><p>AI inference is where the rubber meets the road in machine learning. As AI continues to revolutionize industries, the demand for efficient, scalable inference solutions has skyrocketed.</p><p>Enter Lambda, RunPod, and Together AI - three powerhouses in the AI inference game. Each brings something unique to the table:</p><p>As we dive deeper, I'll unpack how these platforms stack up in real-world scenarios, helping you navigate the complex landscape of AI deployment.</p><h2>Lambda</h2><p><a href="https://lambdalabs.com/?srsltid=AfmBOoreWZZ-9j_rWZwykibeRFSCoZrRAUyinnpcKqBSXeibTmkuiywo">Lambda</a>'s Inference API offers a powerful inference solution, setting itself apart from competitors like RunPod.</p><h3>Infrastructure and hardware</h3><p>- Access to cutting-edge NVIDIA GPUs, including H100s and A100s.</p><p>- Seamless scaling capabilities to handle workloads of any size.</p><p>- Impressively low cold start times, with some models ready in seconds.</p><h3>Software and frameworks</h3><p>- Support for popular models like Llama 3.3 and Qwen 2.5.</p><p>- Seamless integration with PyTorch and TensorFlow.</p><p>- Custom CUDA kernels and optimizations for enhanced performance.</p><p><strong>One of Lambda's standout features is its ability to spin up GPU instances on demand, allowing users to train and deploy ML models without long-term commitments. </strong>This flexibility, combined with competitive pricing and zero hidden fees, makes it an attractive option for both academic institutions and enterprises.</p><h2>RunPod</h2><p><a href="https://www.runpod.io/">RunPod</a> offers a flexible solution for generative AI workloads, challenging competitors like Lambda in the GPU cloud space.</p><h3>Infrastructure and hardware</h3><p>- Diverse GPU options, including NVIDIA A100s and RTX 4090s.</p><p>- On-demand scaling for fluctuating workloads.</p><p>- Impressively low cold start times, often under 30 seconds.</p><h3>Software and frameworks</h3><p>- Support for popular models like Stable Diffusion and GPT-J.</p><p>- Seamless integration with PyTorch, TensorFlow, and other ML libraries.</p><p>- Custom container support for optimized environments.</p><p>The platform excels in running ML models for various applications, from fine-tuning large language models to powering generative AI apps.</p><p><strong>A standout feature is RunPod's job queueing system, allowing users to manage complex AI workloads effectively. </strong>This, combined with competitive pricing and a community cloud approach, makes it attractive for both academic institutions and enterprises.</p><h2>Together AI</h2><p><a href="https://www.together.ai/">Together AI</a>'s Inference Engine offers a compelling alternative to platforms like Lambda and RunPod.</p><h3>Infrastructure and hardware</h3><p>- Access to cutting-edge NVIDIA GPUs, including H200s and GB200s.</p><p>- Impressive scaling capabilities, handling millions of requests per second.</p><p>- Industry-leading cold start times, often under 100ms.</p><h3>Software and frameworks</h3><p>- Support for 200+ open-source AI models, including Llama 3 and Mixtral.</p><p>- Seamless integration with PyTorch and TensorFlow.</p><p>- Custom CUDA kernels, including the Together Kernel Collection (TKC).</p><p><strong>A standout feature is the engine's ability to fine-tune models on-the-fly, allowing for continuous optimization of AI workloads. </strong>This, combined with job queueing and auto-scaling capabilities, makes it an attractive option for both academic institutions and large-scale enterprises.</p><p><strong>With support for private image repositories and a secure cloud environment, Together AI provides a comprehensive solution for AI researchers and engineers looking to push the boundaries of machine learning.</strong></p><h2>Comparative overview</h2><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!hF1x!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F63655e92-c967-4bcc-b6ce-f2f62ffec019_1600x753.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!hF1x!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F63655e92-c967-4bcc-b6ce-f2f62ffec019_1600x753.png 424w, https://substackcdn.com/image/fetch/$s_!hF1x!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F63655e92-c967-4bcc-b6ce-f2f62ffec019_1600x753.png 848w, https://substackcdn.com/image/fetch/$s_!hF1x!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F63655e92-c967-4bcc-b6ce-f2f62ffec019_1600x753.png 1272w, https://substackcdn.com/image/fetch/$s_!hF1x!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F63655e92-c967-4bcc-b6ce-f2f62ffec019_1600x753.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!hF1x!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F63655e92-c967-4bcc-b6ce-f2f62ffec019_1600x753.png" width="1456" height="685" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/63655e92-c967-4bcc-b6ce-f2f62ffec019_1600x753.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:685,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!hF1x!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F63655e92-c967-4bcc-b6ce-f2f62ffec019_1600x753.png 424w, https://substackcdn.com/image/fetch/$s_!hF1x!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F63655e92-c967-4bcc-b6ce-f2f62ffec019_1600x753.png 848w, https://substackcdn.com/image/fetch/$s_!hF1x!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F63655e92-c967-4bcc-b6ce-f2f62ffec019_1600x753.png 1272w, https://substackcdn.com/image/fetch/$s_!hF1x!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F63655e92-c967-4bcc-b6ce-f2f62ffec019_1600x753.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h3>Performance metrics</h3><p>When comparing Lambda, RunPod, and Together AI, performance metrics are crucial for AI researchers and developers looking to deploy and scale AI models efficiently.</p><p><strong>In terms of inference speed, Together AI's proprietary kernels, including optimized MHA and GEMMs, give it an edge, especially for LLM inference</strong>. Their Inference Engine claims to be up to 75% faster than base PyTorch implementations. RunPod, with its globally distributed GPU cloud across 30 regions, offers low-latency inference with sub-250ms cold start times. Lambda Labs, while not providing specific numbers, emphasizes its high-performance infrastructure.</p><p>Throughput capabilities vary, with RunPod handling millions of inference requests daily. <strong>Together AI boasts the ability to process up to 5M tokens per minute for LLMs on their Scale tier</strong>. Lambda Labs doesn't publicly disclose specific throughput numbers, but their focus on high-performance hardware suggests competitive capabilities.</p><p>Memory efficiency is a key factor, especially for large AI models. <strong>Together AI's sparse mixture of experts architecture in models like Mixtral allows for efficient use of GPU memory.</strong> RunPod offers flexible GPU options, including high-memory instances like the 80GB A100, enabling efficient handling of memory-intensive AI tasks.</p><h3>Applications and use cases</h3><h4>NLP</h4><p>All three platforms excel in NLP tasks. Lambda Labs and RunPod provide access to popular open-source models like <a href="https://www.ankursnewsletter.com/p/comparing-open-source-ai-models-llama">LLaMA</a> and GPT-J, enabling text generation and chatbot development. <strong>Together AI offers over 200 open-source AI models, including specialized NLP models for various tasks</strong>.</p><p>For language translation, Together AI's multilingual models like Llama 3.3 70B stand out, supporting translation across multiple languages. RunPod's serverless infrastructure allows for easy deployment of custom translation.</p><p>Sentiment analysis can be efficiently performed on all platforms, with Together AI offering pre-trained models specifically optimized for such tasks.</p><h4>Computer vision</h4><p>Image classification and object detection are well-supported across the board. <strong>RunPod offers templates for popular computer vision frameworks like YOLO, making it easy to deploy these models</strong>. Lambda Labs' high-performance GPUs are particularly suited for compute-intensive vision tasks.</p><p>For image generation, all three platforms support stable diffusion models. RunPod, in particular, has gained popularity in the AI art community for its easy deployment of text-to-image models.</p><p>Video analysis capabilities are available on all platforms, with RunPod's scalable infrastructure being particularly well-suited for handling the high computational demands of video processing.</p><h4>Audio processing</h4><p>Speech recognition and text-to-speech applications are supported across all platforms. Together AI's Inference Engine is optimized for various modalities, including audio processing. RunPod's serverless infrastructure allows for flexible deployment of audio AI models.</p><p>For music generation, all platforms provide the necessary compute resources, with Together AI offering specialized models for audio tasks.</p><h4>Multimodal AI</h4><p>Combined text, image, and audio processing is an area where these platforms truly shine. <strong>Together AI's Llama 3.3 90B Vision model exemplifies their capabilities in multimodal AI. RunPod's flexible infrastructure allows developers to create custom multimodal pipelines.</strong></p><p>Virtual assistants and interactive AI applications can be built and deployed on all three platforms, leveraging their robust infrastructure and support for various AI models.</p><h3>Developer experience</h3><h4>API integration</h4><p>Together AI offers OpenAI-compatible APIs, making it easy for developers familiar with OpenAI's ecosystem to transition. RunPod provides a user-friendly <strong>CLI tool for seamless integration and deployment.</strong> Lambda Labs focuses on providing a straightforward interface for managing GPU instances.</p><p>All three platforms offer comprehensive documentation and support multiple programming languages. RunPod and Together AI provide Python and JavaScript SDKs, while Lambda Labs offers APIs compatible with popular ML frameworks.</p><h4>Customization and fine-tuning</h4><p>Model customization options are available across all platforms. Together AI stands out with its fine-tuning service that allows complete model ownership. <strong>RunPod's support for custom containers enables developers to fine-tune and deploy proprietary models.</strong> Lambda Labs provides the necessary infrastructure for custom model training and deployment.</p><h4>Monitoring and analytics</h4><p>RunPod offers real-time usage analytics for endpoints, execution time analytics, and real-time logs for easy debugging. <strong>Together AI provides a monitoring dashboard with data retention varying from 24 hours to 1 year, depending on the plan</strong>. Lambda Labs offers basic monitoring tools, though specific details are not publicly available.</p><h3>Scalability and production readiness</h3><h4>Auto-scaling features</h4><p>RunPod excels in auto-scaling, with the ability to scale from 0 to hundreds of instances in seconds across multiple regions. Together AI offers scaling up to 9,000 requests per minute on their Scale tier. <strong>Lambda Labs provides scaling capabilities, though specific details are not publicly disclosed.</strong></p><p>All three platforms offer strategies for handling varying workloads and managing concurrent requests efficiently. Cost optimization is a key focus, with RunPod and Together AI offering serverless options to minimize idle costs.</p><h3>Security and compliance</h3><p>Data privacy and security are paramount for all three platforms. RunPod is working towards SOC2, GDPR, and HIPAA compliance. <strong>Together AI ensures SOC 2 and HIPAA compliance. Lambda Labs emphasizes security in their infrastructure, though specific certifications are not publicly listed.</strong></p><p>All platforms offer encryption for data in transit and at rest. Together AI and RunPod provide options for deploying models in secure, isolated environments.</p><h3>Cost analysis</h3><h4>Pricing models</h4><p><a href="https://www.ankursnewsletter.com/p/how-to-calculate-ai-roi-for-your">Pricing models </a>vary across platforms. RunPod offers per-hour pricing for GPU instances, with rates starting as low as $0.67 per hour for an A40 GPU. <strong>Together AI uses a per-token pricing model for inference, with rates varying by model size and type</strong>. Lambda Labs' pricing is not publicly disclosed in the provided search results.</p><p><strong>Hidden costs and additional fees are minimal on RunPod, with zero fees for egress/ingress</strong>. Together AI's pricing is transparent, with no hidden costs mentioned. Lambda Labs' fee structure is not detailed in the available information.</p><h4>Cost optimization strategies</h4><p>All platforms offer strategies for reducing inference costs. RunPod's serverless option allows users to pay only for compute time used. Together AI provides different model variants (Lite, Turbo, Reference) to balance cost and performance.</p><p>Batch processing is supported across all platforms, offering a cost-effective alternative to real-time inference for suitable use cases. The trade-offs between batch processing and real-time inference depend on the specific application requirements and should be carefully considered when optimizing costs.</p><div><hr></div><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!-kh9!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!-kh9!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!-kh9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png" width="1456" height="246" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:246,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:172293,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:&quot;&quot;,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!-kh9!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p>I also host an AI podcast and content series called &#8220;Pioneers.&#8221; This series takes you on an enthralling journey into the minds of AI visionaries, founders, and CEOs who are at the forefront of innovation through AI in their organizations.</p><p>To learn more, please visit <a href="https://www.multimodal.dev/blog">Pioneers</a> on Beehiiv.</p><div><hr></div><h2>Wrapping up</h2><p>Lambda Labs, RunPod, and Together AI each offer unique strengths in the AI infrastructure space. The choice between them depends on specific project requirements, scaling needs, and budget constraints. RunPod stands out for its flexibility and cost-effectiveness, Together AI for its specialized AI offerings and performance optimizations, and Lambda Labs for its focus on high-performance computing.</p><p>I&#8217;ll come back with more such comparisons next week. <br><br>Until then,</p><p>Ankur.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.ankursnewsletter.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">If you liked reading this post, consider subscribing for more AI content.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Comparing Open-Source AI Models: LLaMA 3 vs Qwen 2.5 vs Mixtral]]></title><description><![CDATA[Dive into a comprehensive comparison of 2025's leading open-source AI models: Llama 3, Qwen 2.5, and Mixtral. Discover their architectures, benchmarks, and real-world applications.]]></description><link>https://www.ankursnewsletter.com/p/comparing-open-source-ai-models-llama</link><guid isPermaLink="false">https://www.ankursnewsletter.com/p/comparing-open-source-ai-models-llama</guid><dc:creator><![CDATA[Ankur A. Patel]]></dc:creator><pubDate>Fri, 17 Jan 2025 14:02:05 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!F3y6!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed070b62-f4ba-45fa-9a31-11afb4ab462c_1200x600.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!F3y6!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed070b62-f4ba-45fa-9a31-11afb4ab462c_1200x600.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!F3y6!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed070b62-f4ba-45fa-9a31-11afb4ab462c_1200x600.png 424w, https://substackcdn.com/image/fetch/$s_!F3y6!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed070b62-f4ba-45fa-9a31-11afb4ab462c_1200x600.png 848w, https://substackcdn.com/image/fetch/$s_!F3y6!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed070b62-f4ba-45fa-9a31-11afb4ab462c_1200x600.png 1272w, https://substackcdn.com/image/fetch/$s_!F3y6!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed070b62-f4ba-45fa-9a31-11afb4ab462c_1200x600.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!F3y6!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed070b62-f4ba-45fa-9a31-11afb4ab462c_1200x600.png" width="1200" height="600" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/ed070b62-f4ba-45fa-9a31-11afb4ab462c_1200x600.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:600,&quot;width&quot;:1200,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:390090,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!F3y6!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed070b62-f4ba-45fa-9a31-11afb4ab462c_1200x600.png 424w, https://substackcdn.com/image/fetch/$s_!F3y6!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed070b62-f4ba-45fa-9a31-11afb4ab462c_1200x600.png 848w, https://substackcdn.com/image/fetch/$s_!F3y6!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed070b62-f4ba-45fa-9a31-11afb4ab462c_1200x600.png 1272w, https://substackcdn.com/image/fetch/$s_!F3y6!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fed070b62-f4ba-45fa-9a31-11afb4ab462c_1200x600.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.ankursnewsletter.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Subscribe today to join our community of 1830 AI enthusiasts and business leaders for more insider takes and practical advice on implementing AI in business.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h2>Key Takeaways</h2><p>1. <strong>Llama 3, Qwen 2.5, and Mixtral</strong> represent the current leaders in open-source language models, with Qwen 2.5 72B slightly edging out competitors on MMLU benchmarks at 86.1%.</p><p>2. Each model offers <strong>unique deployment advantages</strong>: Llama 3 features 15% more efficient tokenization, Qwen 2.5 provides flexible model sizes from 0.5B to 72B, and Mixtral achieves 6x faster inference through its sparse mixture of experts architecture.</p><p>3. For <strong>specialized capabilities</strong>, Llama 3 excels in multimodal tasks, Qwen 2.5 dominates structured data handling, and Mixtral shines in multilingual support and mathematical reasoning.</p><p>4. In <strong>technical evaluations</strong>, Llama 3.1 leads in HumanEval code generation at 80.5%, while Qwen 2.5 demonstrates superior performance in mathematical reasoning with 83.1% on MATH benchmarks.</p><p>5. <strong>Future developments</strong> include Meta pushing beyond 405B parameters for Llama models and Alibaba expanding Qwen's capabilities in tool use and agentic applications, while both focus on enhanced security features.</p><div><hr></div><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!R9ua!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!R9ua!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!R9ua!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png" width="1456" height="246" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/bea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:246,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:273651,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:&quot;&quot;,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!R9ua!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1456w" sizes="100vw" loading="lazy" fetchpriority="high"></picture><div></div></div></a></figure></div><p>Last year, I started Multimodal, a Generative AI company that helps organizations automate complex, knowledge-based workflows using AI Agents. Check it out <a href="https://www.multimodal.dev/">here</a>.</p><div><hr></div><p>The landscape of <a href="https://www.ankursnewsletter.com/publish/posts/detail/136281974?referrer=%2Fpublish%2Fposts%3Fsearch%3Dopen%2520sour">open-source large language models</a> has dramatically evolved in the past year, with three foundation models emerging as clear leaders: <strong><a href="https://www.ankursnewsletter.com/publish/posts/detail/135591212?referrer=%2Fpublish%2Fposts%3Fsearch%3Dopen%2520sour">Meta's Llama 3</a>, Alibaba's Qwen 2.5, and <a href="https://www.ankursnewsletter.com/publish/posts/detail/140574582?referrer=%2Fpublish%2Fposts%3Fsearch%3Dopen%2520sour">Mixtral</a>'s sparse mixture of experts architecture.</strong> These models represent a significant leap forward in performance, efficiency, and real-world applications.</p><p>For enterprises and developers looking to leverage open weights models, understanding the nuances between these architectures is crucial. Let's dive deep into how these competing models stack up against each other across various dimensions.</p><h2>Model architectures</h2><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!gmoI!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F20b0b48e-91cd-4392-b1be-0f7df5cae930_1600x490.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!gmoI!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F20b0b48e-91cd-4392-b1be-0f7df5cae930_1600x490.png 424w, https://substackcdn.com/image/fetch/$s_!gmoI!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F20b0b48e-91cd-4392-b1be-0f7df5cae930_1600x490.png 848w, https://substackcdn.com/image/fetch/$s_!gmoI!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F20b0b48e-91cd-4392-b1be-0f7df5cae930_1600x490.png 1272w, https://substackcdn.com/image/fetch/$s_!gmoI!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F20b0b48e-91cd-4392-b1be-0f7df5cae930_1600x490.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!gmoI!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F20b0b48e-91cd-4392-b1be-0f7df5cae930_1600x490.png" width="1456" height="446" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/20b0b48e-91cd-4392-b1be-0f7df5cae930_1600x490.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:446,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!gmoI!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F20b0b48e-91cd-4392-b1be-0f7df5cae930_1600x490.png 424w, https://substackcdn.com/image/fetch/$s_!gmoI!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F20b0b48e-91cd-4392-b1be-0f7df5cae930_1600x490.png 848w, https://substackcdn.com/image/fetch/$s_!gmoI!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F20b0b48e-91cd-4392-b1be-0f7df5cae930_1600x490.png 1272w, https://substackcdn.com/image/fetch/$s_!gmoI!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F20b0b48e-91cd-4392-b1be-0f7df5cae930_1600x490.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h3>Llama 3</h3><p>- Employs a decoder-only transformer architecture with advanced grouped-query attention mechanism.</p><p>- Introduces a new tokenizer with <strong>128K vocabulary</strong>, enabling more efficient text processing.</p><p>- Utilizes RoPE embeddings and sliding window attention for enhanced performance.</p><p>- Features specialized instruction tuning and direct preference optimization for improved output quality.</p><h3>Qwen 2.5</h3><p>- Built on a dense <strong>decoder-only architecture</strong> with RoPE, SwiGLU, and RMSNorm components.</p><p>- Implements attention QKV bias and tied word embeddings for better performance.</p><p>- Supports extensive multilingual capabilities across 29 languages.</p><p>- Incorporates YaRN for efficient context window extension.</p><h3>Mixtral</h3><p>- Features innovative sparse mixture-of-experts (MoE) architecture.</p><p>- Employs 8 <strong>expert networks with top-2 routing per layer</strong>.</p><p>- Shares attention parameters while varying feed-forward blocks.</p><p>- Uses byte-fallback BPE tokenizer for robust character handling.</p><h2>Parameter scaling &amp; efficiency</h2><h3>Llama 3 Series</h3><p>- Scales from 1B to 405B parameters across different variants.</p><p>- 405B model trained using 16,000 H100 GPUs.</p><p>- Achieves <strong>95% training efficiency</strong> through advanced error detection.</p><p>- Supports efficient quantization from BF16 to FP8 for deployment.</p><h3>Qwen 2.5</h3><p>- Ranges from 0.5B to 72B parameters with specialized variants.</p><p>- Trained on 18 trillion tokens of diverse data.</p><p>- Optimized for both <strong>high-end and edge device deployment</strong>.</p><p>- Features dedicated math and coding variants for specialized tasks.</p><h3>Mixtral</h3><p>- Total parameter count of 46.7B with only 12.9B active during inference.</p><p>- Achieves computation efficiency equivalent to a 13B parameter model.</p><p>- Requires <strong>2x sequence length operations for expert routing</strong>.</p><p>- Optimized for both training and inference efficiency.</p><h2>Training &amp; Fine-tuning</h2><h3>Llama 3 Series</h3><p>- Leverages <strong>supervised <a href="https://www.multimodal.dev/post/understanding-fine-tuning-in-deep-learning">fine-tuning</a></strong> and direct preference optimization.</p><p>- Implements extensive safety evaluations during training.</p><p>- Uses advanced scaling laws for optimal data mixing.</p><p>- Incorporates multi-stage instruction tuning process.</p><h3>Qwen 2.5</h3><p>- Employs specialized training for code and mathematical tasks.</p><p>- Features <strong>comprehensive instruction</strong> tuning across domains.</p><p>- Supports structured data handling and JSON generation.</p><p>- Includes extensive security and safety evaluations.</p><h3>Mixtral</h3><p>- Utilizes expert-specific training procedures.</p><p>- Implements sliding window attention with 8K context training.</p><p>- Features <strong>grouped-query attention</strong> for faster inference.</p><p>- Maintains balanced expert utilization through router network.</p><h2>General knowledge performance</h2><h3>MMLU benchmark results</h3><p>- Qwen 2.5 72B leads marginally at 86.1% on MMLU.</p><p>- Llama 3.1 70B follows closely at 86.0%.</p><p>- Mixtral maintains GPT-3.5 competitive scores at 83.7%.</p><h3>Specialized knowledge areas</h3><p>- Qwen 2.5: Excels in mathematical reasoning (83.1% on MATH).</p><p>- Llama 3.1: Superior in world knowledge tasks (81.2%).</p><p>- Mixtral: Strong multilingual performance across 30+ languages.</p><h3>Technical capabilities</h3><h4>Code generation</h4><p>- Llama 3.1: 80.5% on HumanEval after instruction tuning.</p><p>- Qwen 2.5: 79.8% on HumanEval, 82.3% on MBPP.</p><p>- Mixtral: 78.2% on HumanEval with efficient inference.</p><h4>Mathematical reasoning</h4><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!FZyc!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c6aa585-9db1-4c63-9be2-9aebec640691_962x261.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!FZyc!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c6aa585-9db1-4c63-9be2-9aebec640691_962x261.png 424w, https://substackcdn.com/image/fetch/$s_!FZyc!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c6aa585-9db1-4c63-9be2-9aebec640691_962x261.png 848w, https://substackcdn.com/image/fetch/$s_!FZyc!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c6aa585-9db1-4c63-9be2-9aebec640691_962x261.png 1272w, https://substackcdn.com/image/fetch/$s_!FZyc!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c6aa585-9db1-4c63-9be2-9aebec640691_962x261.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!FZyc!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c6aa585-9db1-4c63-9be2-9aebec640691_962x261.png" width="962" height="261" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/3c6aa585-9db1-4c63-9be2-9aebec640691_962x261.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:261,&quot;width&quot;:962,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!FZyc!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c6aa585-9db1-4c63-9be2-9aebec640691_962x261.png 424w, https://substackcdn.com/image/fetch/$s_!FZyc!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c6aa585-9db1-4c63-9be2-9aebec640691_962x261.png 848w, https://substackcdn.com/image/fetch/$s_!FZyc!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c6aa585-9db1-4c63-9be2-9aebec640691_962x261.png 1272w, https://substackcdn.com/image/fetch/$s_!FZyc!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c6aa585-9db1-4c63-9be2-9aebec640691_962x261.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h2>Real-world applications</h2><h4>Tool use &amp; agents</h4><p>- Llama 3.1 shows superior performance in agentic applications.</p><p>- Qwen 2.5 excels in structured data handling.</p><p>- Mixtral demonstrates efficient scaling across diverse tasks.</p><h2>Deployment &amp; efficiency</h2><p>Let's talk about what really matters when you're putting these models into production. Each one brings something unique to the table, and I'll break down why that matters for your real-world deployments.</p><h3>Llama 3's optimization story</h3><p>The latest Llama brings some impressive efficiency gains to the table. That 15% more efficient tokenizer isn't just a number - it means you can process more text with less compute. Plus, if you're working with edge devices, you'll love how it handles FP8 quantization, pushing 1.4x better throughput in production.</p><h3>Qwen 2.5's flexibility play</h3><p>Here's where Qwen 2.5 really shines - it's like having a Swiss Army knife of models. Need something light for a simple task? Grab the 0.5B variant. Going all-out on performance? The 72B version has you covered. The best part? Its permissive license means you can actually build real products without legal headaches.</p><h3>Mixtral's speed</h3><p>Now this is where things get interesting. Mixtral's sparse mixture of experts approach isn't just clever architecture - it's <strong>6x faster inference in the real world</strong>. You're only using two experts per token, but getting performance that keeps up with the big players. For production environments where every millisecond counts, this is game-changing stuff.</p><p>The beauty of these open models is how they're pushing the envelope on what's possible with efficient inference. Whether you're optimizing for speed, flexibility, or resource usage, there's a clear path forward.</p><h2>Specialized capabilities</h2><h3>Llama 3's multimodal magic</h3><p>Here's where Llama 3 gets really interesting - it's not just about text anymore. The model can actually understand images alongside text, making it a game-changer for tasks like visual analysis and image-based conversations. If you're building applications that need to work with both text and visuals, this is huge.</p><h3>Qwen 2.5's structured data superpower</h3><p>Want to talk about handling structured data? Qwen 2.5 absolutely crushes it. It's like having a data analyst built into your model - <strong>particularly impressive with JSON and complex data structures</strong>. For enterprise applications where clean, structured outputs matter, this is your go-to choice.</p><h3>Mixtral's multilingual mastery</h3><p>Now this is where Mixtral really shows off. Not only does it handle multiple languages like a champ, but it's also <strong>surprisingly good at mathematical reasoning.</strong> The cool part? It does all this while being super efficient with its sparse mixture of experts. You're getting top-tier performance across languages and technical tasks without breaking the bank on compute resources.</p><h2>Enterprise applications</h2><p>Let's talk about where these models really shine in the enterprise and what's coming down the pipeline.</p><p><strong>If you're handling large-scale deployments</strong>, Llama 3's efficient inference and strong performance make it your go-to choice. The model's supervised fine-tuning and direct preference optimization really show in real-world scenarios.</p><p><strong>For those watching their budget</strong>, Qwen 2.5's high-quality sparse mixture architecture delivers impressive results while keeping costs in check. Its permissive license and open weights make it particularly attractive for companies needing customization options.</p><p><strong>When it comes to development work</strong>, these models are absolute powerhouses. Qwen 2.5 crushes it in code generation with HumanEval scores over 85%, while Llama 3's latest version brings some serious improvements to API integration and technical documentation.</p><div><hr></div><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!-kh9!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!-kh9!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!-kh9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png" width="1456" height="246" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:246,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:172293,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:&quot;&quot;,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!-kh9!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p>I also host an AI podcast and content series called &#8220;Pioneers.&#8221; This series takes you on an enthralling journey into the minds of AI visionaries, founders, and CEOs who are at the forefront of innovation through AI in their organizations.</p><p>To learn more, please visit <a href="https://www.multimodal.dev/blog">Pioneers</a> on Beehiiv.</p><div><hr></div><h2>Wrapping up</h2><p>The future's looking pretty exciting. Meta's cooking up some bigger Llama models with parameters pushing past 405B, and they're focusing on multimodal capabilities and longer context windows. Not to be outdone, Alibaba's working on expanding Qwen's abilities with better tool use and agentic applications</p><p>The open source community is driving faster innovation in model architectures too. <strong>We're seeing new approaches to a mixture of experts and query attention mechanisms that could really change the game.</strong> Plus, both teams are working on better security features and addressing potential risks - super important for enterprise adoption.</p><p>The best part? These competing models are pushing each other to get better and better, which means we all win. Whether you're building the next big thing or just need solid language model performance, there's never been a better time to jump in.<br><br>I&#8217;ll come back next week with more on AI models and enterprise deployment.</p><p>Until then,</p><p>Ankur.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.ankursnewsletter.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">If you liked reading this post, consider subscribing for more AI content.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Tier 1 Vs Tier 2 AI Players: A Comparison]]></title><description><![CDATA[Explore groundbreaking advancements from Tier 1 giants like OpenAI & Google, and innovative solutions from Tier 2 players like Groq & Cerebras. Discover which AI powerhouse best suits your needs.]]></description><link>https://www.ankursnewsletter.com/p/tier-1-vs-tier-2-ai-players-a-comparison</link><guid isPermaLink="false">https://www.ankursnewsletter.com/p/tier-1-vs-tier-2-ai-players-a-comparison</guid><dc:creator><![CDATA[Ankur A. Patel]]></dc:creator><pubDate>Thu, 09 Jan 2025 16:31:15 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!_h6r!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c39b28d-b723-424b-ab44-2ccbddb8e362_1200x600.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!_h6r!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c39b28d-b723-424b-ab44-2ccbddb8e362_1200x600.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!_h6r!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c39b28d-b723-424b-ab44-2ccbddb8e362_1200x600.png 424w, https://substackcdn.com/image/fetch/$s_!_h6r!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c39b28d-b723-424b-ab44-2ccbddb8e362_1200x600.png 848w, https://substackcdn.com/image/fetch/$s_!_h6r!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c39b28d-b723-424b-ab44-2ccbddb8e362_1200x600.png 1272w, https://substackcdn.com/image/fetch/$s_!_h6r!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c39b28d-b723-424b-ab44-2ccbddb8e362_1200x600.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!_h6r!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c39b28d-b723-424b-ab44-2ccbddb8e362_1200x600.png" width="1200" height="600" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/3c39b28d-b723-424b-ab44-2ccbddb8e362_1200x600.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:600,&quot;width&quot;:1200,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:370103,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!_h6r!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c39b28d-b723-424b-ab44-2ccbddb8e362_1200x600.png 424w, https://substackcdn.com/image/fetch/$s_!_h6r!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c39b28d-b723-424b-ab44-2ccbddb8e362_1200x600.png 848w, https://substackcdn.com/image/fetch/$s_!_h6r!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c39b28d-b723-424b-ab44-2ccbddb8e362_1200x600.png 1272w, https://substackcdn.com/image/fetch/$s_!_h6r!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c39b28d-b723-424b-ab44-2ccbddb8e362_1200x600.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.ankursnewsletter.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Subscribe today to join our community of 1820 AI enthusiasts and business leaders for more insider takes and practical advice on implementing AI in business.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h2>Key Takeaways</h2><ul><li><p>Tier 1 companies like <strong>OpenAI, Google, and Anthropic</strong> are leading in developing advanced AI models with groundbreaking capabilities like multimodal processing and agentic behavior.</p></li><li><p>Tier 2 companies like <strong>Groq and Cerebras</strong> are innovating in AI hardware, offering solutions like high-speed chips and wafer-scale engines for improved performance and efficiency.</p></li><li><p>AI21 Labs specializes in <strong>task-specific AI models</strong>, delivering high accuracy and efficiency for specific applications like content generation and summarization.</p></li><li><p>The AI landscape is characterized by <strong>rapid evolution, focusing on developing more powerful, efficient, and ethical AI systems</strong>.</p></li></ul><div><hr></div><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!R9ua!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!R9ua!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!R9ua!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png" width="1456" height="246" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/bea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:246,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:273651,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:&quot;&quot;,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!R9ua!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1456w" sizes="100vw" loading="lazy" fetchpriority="high"></picture><div></div></div></a></figure></div><p>Last year, I started Multimodal, a Generative AI company that helps organizations automate complex, knowledge-based workflows using AI Agents. Check it out <a href="https://www.multimodal.dev/">here</a>.</p><div><hr></div><p>The AI landscape is evolving at breakneck speed, with agentic AI emerging as a game-changer. These AI systems can autonomously perform tasks, make decisions, and interact with their environment. This opens up several applications in knowledge-based industries. Also, as the field matures, a clear hierarchy has emerged among AI developers.</p><p><strong>Tier One giants like OpenAI, Google, Anthropic, and Meta are pushing the boundaries with massive models and groundbreaking capabilities.</strong> Meanwhile, Tier Two contenders such as Groq, Cerebras, and AI21 Labs are carving out niches with specialized solutions and innovative approaches, challenging the status quo and driving the industry forward.</p><p>Let&#8217;s talk about what each of these players bring to the table, and which of them will suit your applications the best.</p><h2>Tier 1 Players</h2><h3>Open AI</h3><p><a href="https://openai.com/">OpenAI</a>'s GPT-4o ("o" for "omni") represents a significant leap in multimodal AI capabilities. This model can process and generate text, audio, images, and video inputs with remarkable speed and accuracy. Key advancements include:</p><p>- <strong>Real-time audio response</strong>: GPT-4o can respond to audio inputs in as little as 232 milliseconds, with an average of 320 milliseconds, comparable to human conversation speed.</p><p>- <strong>Enhanced vision capabilities</strong>: The model excels at understanding and discussing images, outperforming previous iterations in visual perception benchmarks.</p><p>- <strong>Multilingual proficiency</strong>: GPT-4o supports over 50 languages, with improved efficiency in processing non-Roman scripts.</p><p>GPT-4o mini, a more cost-efficient variant, offers similar capabilities at a fraction of the cost. It outperforms GPT-3.5 Turbo while being 60% cheaper.</p><h4>Application</h4><p>- Real-time translation and interpretation.</p><p>- Advanced data analysis and coding assistance.</p><p>- Emotional recognition and expression in interactions.</p><p>- Accessibility tools for visually impaired users.</p><h4>Future developments</h4><p>- Enhanced multimodal reasoning capabilities.</p><p>- Improved safety measures and alignment with human values.</p><p>- Integration with various APIs for complex workflow automation.</p><h3>Google (Gemini 2.0)</h3><p><a href="https://blog.google/technology/google-deepmind/google-gemini-ai-update-december-2024/">Google's Gemini 2.0</a> represents a significant leap in multimodal AI capabilities, designed for the "agentic era" of artificial intelligence. Key features and capabilities include:</p><p>- <strong>Enhanced multimodal processing</strong>: Gemini 2.0 can understand and generate content across text, images, audio, and video modalities.</p><p>- <strong>Native tool use</strong>: The model can natively call tools like Google Search, Maps, and code execution, as well as integrate with third-party functions.</p><p>- <strong>Improved performance</strong>: Gemini 2.0 Flash outperforms its predecessor Gemini 1.5 Pro on key benchmarks while operating at twice the speed.</p><p>- <strong>Real-time interactions</strong>: The Multimodal Live API enables developers to build applications with real-time audio and video streaming capabilities.</p><h4>Applications</h4><p>1. <strong>Advanced search and information retrieval</strong>: Enhancing Google Search with AI Overviews that can handle complex, multi-step queries.</p><p>2. <strong>Personalized education</strong>: Developing tailored lessons and adaptive learning experiences.</p><p>3.<strong> Scientific research assistance</strong>: Summarizing trends and synthesizing information from vast literature.</p><p>4. <strong>Multimodal customer support</strong>: Analyzing product images and providing actionable responses.</p><p>5. <strong>Creative content generation:</strong> Producing integrated responses including text, audio, and images through a single API call.</p><p>Google is rapidly integrating Gemini 2.0 across its ecosystem, from Search to Workspace applications. The model's agentic capabilities are being showcased through prototypes like Project Astra (a universal AI assistant) and Project Mariner (an AI-powered web browsing agent).</p><h4>Future developments</h4><p>- Further enhancements to agentic AI capabilities.</p><p>- Expanded integration across Google's product suite.</p><p>- Continued improvements in multimodal understanding and generation.</p><h3>Meta (Llama 3.2 and 3.3)</h3><p><a href="https://ai.meta.com/blog/meta-llama-3-1/">Meta</a>'s latest Llama models represent significant advancements in open-source AI, pushing the boundaries of performance and accessibility. Key features and capabilities include:</p><p>- <strong>Multimodal support</strong>: Llama 3.2 introduces vision capabilities in its 11B and 90B parameter versions, enabling image understanding alongside text processing.</p><p>- <strong>Enhanced performance</strong>: Llama 3.3 70B offers comprehensive training results and robust understanding across diverse tasks.</p><p>- <strong>Open-source approach</strong>: Meta continues its commitment to democratizing AI research by making these models openly available.</p><p>- <strong>Improved context handling</strong>: Both models feature expanded context windows, allowing for more coherent long-form interactions.</p><h4>Applications</h4><p>1. <strong>Advanced language understanding</strong>: Excelling in tasks like sentiment analysis, text classification, and named entity recognition.</p><p>2. <strong>Multimodal content analysis</strong>: Llama 3.2's vision capabilities enable applications in image captioning and visual question answering.</p><p>3. <strong>Code generation and analysis</strong>: Enhanced performance in programming-related tasks, benefiting software development workflows.</p><p>4. <strong>Creative writing assistance</strong>: Improved language generation for content creation and storytelling.</p><p>5. <strong>Research and academia</strong>: Open-source nature facilitates further innovation and study in the AI community.</p><h4>Future developments</h4><p>- Further improvements in multimodal capabilities.</p><p>- Expanded language support and cross-lingual understanding.</p><p>- Enhanced safety measures and bias mitigation techniques.</p><h3>Claude 3.5 Sonnet</h3><p><a href="https://www.anthropic.com/news/claude-3-5-sonnet">Claude 3.5 Sonnet</a>, Anthropic's latest flagship model, introduces significant improvements and groundbreaking features:</p><p>- Outperforms its predecessor across various benchmarks, particularly in coding tasks.</p><p>- Improves SWE-bench Verified score from <strong>33.4% to 49%, surpassing all publicly available models</strong>.</p><p>- Advances in <strong>TAU-bench</strong>, an agentic tool use task, with gains in retail (69.2%) and airline (46%) domains.</p><p>- First frontier AI model to offer computer use capabilities in a public beta.</p><p>- Interprets screenshots of <strong>graphical user interfaces (GUIs)</strong> and generates appropriate tool calls.</p><p>- Enables navigation of websites, interaction with user interfaces, and completion of complex multi-step processes.</p><h4>Applications</h4><p>- Software development: Assists across the entire lifecycle, from design to maintenance.</p><p>- Data analysis: Extracts insights from visuals like charts and diagrams.</p><p>- Automation: Handles repetitive tasks and operations with increased efficiency.</p><h3>Claude 3.5 Haiku</h3><p><a href="https://www.anthropic.com/claude/haiku">Claude 3.5 Haiku</a>, Anthropic's next-generation fast model, offers impressive capabilities:</p><p>- Matches or surpasses Claude 3 Opus (previously Anthropic's largest model) on many intelligence benchmarks.</p><p>- Achieves similar speed to Claude 3 Haiku while improving across every skill set.</p><p>- Scores 40.6% on SWE-bench Verified, outperforming many agents using state-of-the-art models.</p><p>- 200,000 token context window.</p><p>- Maximum output of 8,192 tokens.</p><p>- Knowledge cut-off date of July 2024.</p><h4>Applications</h4><p>- Real-time content moderation.</p><p>- Fast and accurate code suggestions.</p><p>- Highly interactive chatbots for customer service.</p><p>- Personalized experiences from large datasets (e.g., purchase history analysis).</p><h3>Groq</h3><p><a href="https://groq.com/">Groq</a>'s innovative hardware architecture, centered around its Language Processing Unit (LPU), represents a significant leap in AI inference technology. The LPU is designed with a simplified, deterministic architecture that eliminates the need for complex control circuitry and caches found in traditional processors.</p><p>This streamlined design allows for exceptional speed and efficiency in AI inference tasks. Groq's LPU has demonstrated remarkable performance, achieving up to 814 tokens per second when running models like Gemma 7B. This speed is significantly faster than competing solutions, often outperforming them by 5-15x.</p><p>1. <strong>Natural language processing</strong> for instant language translation and transcription.</p><p>2. <strong>Computer vision</strong> for autonomous vehicles and robotics.</p><p>3. <strong>Financial services for real-time trading</strong> and risk analysis.</p><p>The potential impact of Groq's technology on AI deployment is substantial. Its high-speed, low-latency performance enables more responsive AI applications and opens up new possibilities for AI integration in time-sensitive domains. Additionally, Groq's focus on energy efficiency and cost-effectiveness could make advanced AI capabilities more accessible to a broader range of organizations.</p><h2>Tier 2 Players</h2><h3>Cerebras</h3><p><a href="https://cerebras.ai/">Cerebras</a> has revolutionized AI hardware with its innovative wafer-scale engine (WSE) technology, offering unprecedented performance and scalability for AI training and inference tasks.</p><h4>Wafer-Scale Engine Technology</h4><p>Cerebras' WSE is the largest chip ever built, featuring:</p><p>- 46,225 mm&#178; of silicon, 56x larger than the largest GPU.</p><p>- 900,000 AI-optimized cores in the latest WSE-3.</p><p>- 44 GB of on-chip SRAM memory.</p><p>- 21 petabytes/second of memory bandwidth.</p><p>- Eliminates the need for complex distributed computing across multiple smaller chips.</p><p>- Provides native support for sparse computation, boosting efficiency for AI workloads.</p><p>- Enables training of models with up to 24 trillion parameters on a single system.</p><h4>Benefits</h4><p>- Reduced training time from weeks to days or hours.</p><p>- Lower latency for real-time AI applications.</p><p>- Improved cost-efficiency, with up to 100x better price-performance than GPU solutions.</p><h4>Applications</h4><p>- <strong>Molecular dynamics simulations</strong>: Achieved over 1.1 million simulation steps per second, 748x faster than the world's leading supercomputer.</p><p>- <strong>Drug discovery</strong>: GSK researchers used Cerebras to train complex epigenomic models in 2.5 days instead of 24 days on GPU clusters.</p><p>- <strong>Healthcare</strong>: Mayo Clinic is developing multimodal large language models to improve patient outcomes and diagnoses.</p><p>- <strong>Scientific computing</strong>: Accelerating computational fluid dynamics, AI-augmented modeling, and simulation workloads.</p><p>Cerebras continues to push the boundaries of AI hardware:</p><p>- The CS-3 system, powered by the WSE-3, delivers 125 petaflops of AI performance.</p><p>- Cerebras Wafer-Scale Cluster technology enables near-linear scaling from one to hundreds of nodes.</p><p>- MemoryX technology allows for training models larger than the on-chip memory capacity.</p><h3>AI21 Labs</h3><p><a href="https://www.ai21.com/">AI21 Labs</a> has emerged as a pioneer in the field of natural language processing, with a distinct focus on developing task-specific models that bridge the gap between cutting-edge research and practical enterprise applications.</p><h4>Task-specific models</h4><p>AI21's approach to task-specific models (TSMs) sets them apart in the AI landscape:</p><p>- T<strong>SMs are designed to excel at particular tasks, offering higher accuracy and efficiency compared to general-purpose foundation models.</strong></p><p>- These models deliver out-of-the-box value, cost-effectiveness, and improved accuracy for common commercial tasks.</p><p>- AI21's TSMs include Contextual Answers, Summarize, Paraphrase, and Grammatical Error Correction, available through platforms like Amazon SageMaker JumpStart.</p><p>A key advantage of AI21's TSMs is their ability to refuse answers when questions fall outside their intended context, reducing the risk of hallucinations and improving reliability.</p><h4>Jamba Model Family</h4><p>The Jamba model family represents AI21's latest advancement in language models:</p><p>- <strong>Jamba 1.5 Large</strong> (94B active/398B total parameters) and Jamba 1.5 Mini (12B active/52B total parameters) are state-of-the-art hybrid SSM-Transformer models.</p><p>- They feature a 256K token context window, the longest among open models.</p><p>- The models utilize a novel architecture combining Transformer and Mamba layers, optimizing for both quality and efficiency.</p><p>- Jamba models outperform competitors in speed, with Jamba 1.5 Large demonstrating up to 2.5x faster inference on long contexts.</p><h4>Applications</h4><p>- Content generation and summarization for marketing and journalism.</p><p>- Legal document analysis and e-discovery.</p><p>- Financial report condensation and data extraction.</p><p>- Customer support automation and chatbot development.</p><p>- Academic research assistance and personalized tutoring.</p><p>- The company offers AI21 Studio, a developer platform for building custom text-based AI applications.</p><p>- AI21 has partnered with major cloud providers like <a href="https://www.ankursnewsletter.com/p/aws-bedrock-vs-google-vertex-ai-vs">AWS, Google Cloud, and Microsoft Azure</a> to ensure easy deployment of their models in enterprise environments.</p><p>- Their models are designed for both research purposes and commercial applications, with options for on-premises deployment for industries handling sensitive data.</p><h2>Comparative analysis</h2><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!hIox!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4f343a7e-3a92-4d7e-b59a-05129b03ba8b_1524x536.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!hIox!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4f343a7e-3a92-4d7e-b59a-05129b03ba8b_1524x536.png 424w, https://substackcdn.com/image/fetch/$s_!hIox!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4f343a7e-3a92-4d7e-b59a-05129b03ba8b_1524x536.png 848w, https://substackcdn.com/image/fetch/$s_!hIox!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4f343a7e-3a92-4d7e-b59a-05129b03ba8b_1524x536.png 1272w, https://substackcdn.com/image/fetch/$s_!hIox!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4f343a7e-3a92-4d7e-b59a-05129b03ba8b_1524x536.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!hIox!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4f343a7e-3a92-4d7e-b59a-05129b03ba8b_1524x536.png" width="1456" height="512" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/4f343a7e-3a92-4d7e-b59a-05129b03ba8b_1524x536.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:512,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!hIox!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4f343a7e-3a92-4d7e-b59a-05129b03ba8b_1524x536.png 424w, https://substackcdn.com/image/fetch/$s_!hIox!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4f343a7e-3a92-4d7e-b59a-05129b03ba8b_1524x536.png 848w, https://substackcdn.com/image/fetch/$s_!hIox!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4f343a7e-3a92-4d7e-b59a-05129b03ba8b_1524x536.png 1272w, https://substackcdn.com/image/fetch/$s_!hIox!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4f343a7e-3a92-4d7e-b59a-05129b03ba8b_1524x536.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h4>Strengths and unique selling points</h4><h5>Tier one players</h5><p>- <strong>OpenAI</strong>: Pioneering research in AGI, strong focus on AI safety.</p><p>- <strong>Google</strong>: Vast data resources, integration across multiple platforms.</p><p>- <strong>Anthropic</strong>: Constitutional AI approach, emphasis on ethical AI development.</p><p>- <strong>Meta</strong>: Open-source strategy, large-scale language models.</p><h5>Tier two players</h5><p>- <strong>Groq</strong>: High-performance AI chips, focus on speed and efficiency.</p><p>- <strong>Cerebras</strong>: Wafer-scale engine technology, specialized for AI workloads.</p><p>- <strong>AI21 Labs</strong>: Task-specific models, emphasis on natural language understanding.</p><h4>Competitive advantages</h4><p>OpenAI and Google lead in general-purpose AI models, with GPT-4 and Gemini 2.0 setting industry benchmarks. <strong>Anthropic differentiates itself through its focus on AI safety and ethics, while Meta leverages its massive user base to train and deploy AI models.</strong></p><p>Groq and Cerebras compete in the AI hardware space, offering unique solutions for high-performance computing. AI21 Labs carves out a niche with its focus on specialized language models for specific tasks.</p><h4>Technological differentiators</h4><p>- <strong>OpenAI</strong>: Advanced multimodal capabilities in GPT-4o</p><p>- <strong>Google</strong>: Native tool use and multimodal processing in Gemini 2.0</p><p>- <strong>Anthropic</strong>: Constitutional AI for safer, more controllable AI systems</p><p>- <strong>Meta</strong>: Open-source approach with Llama models</p><p>- <strong>Groq</strong>: RealScale&#8482; chip-to-chip interconnect technology</p><p>- <strong>Cerebras</strong>: Wafer-scale engine for massive parallel processing</p><p>- <strong>AI21 Labs</strong>: Jamba model family with task-specific optimizations</p><div><hr></div><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!-kh9!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!-kh9!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!-kh9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png" width="1456" height="246" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:246,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:172293,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:&quot;&quot;,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!-kh9!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p>I also host an AI podcast and content series called &#8220;Pioneers.&#8221; This series takes you on an enthralling journey into the minds of AI visionaries, founders, and CEOs who are at the forefront of innovation through AI in their organizations.</p><p>To learn more, please visit <a href="https://www.multimodal.dev/blog">Pioneers</a> on Beehiiv.</p><div><hr></div><h2>Wrapping up</h2><p>As we look to the future of AI, it's clear that both Tier One and Tier Two players will continue to shape the landscape in profound ways. <strong>The giants like OpenAI, Google, Anthropic, and Meta are pushing the boundaries of general-purpose AI, while specialized players such as Groq, Cerebras, and AI21 Labs are carving out crucial niches with their innovative approaches.</strong></p><p>As the field evolves, we can expect to see increased collaboration between these players. The key to success will lie in balancing cutting-edge capabilities with ethical considerations and real-world applicability.</p><p>I&#8217;ll come back next week with more insightful comparisons and analysis.</p><p>Until then,</p><p>Ankur.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.ankursnewsletter.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">If you liked reading this post, consider subscribing for more AI content.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Starting 2025: A new year of AI breakthroughs]]></title><description><![CDATA[Welcoming 2025 and exploring the future of AI]]></description><link>https://www.ankursnewsletter.com/p/starting-2025-a-new-year-of-ai-breakthroughs</link><guid isPermaLink="false">https://www.ankursnewsletter.com/p/starting-2025-a-new-year-of-ai-breakthroughs</guid><dc:creator><![CDATA[Ankur A. Patel]]></dc:creator><pubDate>Thu, 02 Jan 2025 14:01:50 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!-e6K!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6102779b-f9ab-4602-9a57-23009be33879_1200x600.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!-e6K!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6102779b-f9ab-4602-9a57-23009be33879_1200x600.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!-e6K!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6102779b-f9ab-4602-9a57-23009be33879_1200x600.png 424w, https://substackcdn.com/image/fetch/$s_!-e6K!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6102779b-f9ab-4602-9a57-23009be33879_1200x600.png 848w, https://substackcdn.com/image/fetch/$s_!-e6K!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6102779b-f9ab-4602-9a57-23009be33879_1200x600.png 1272w, https://substackcdn.com/image/fetch/$s_!-e6K!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6102779b-f9ab-4602-9a57-23009be33879_1200x600.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!-e6K!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6102779b-f9ab-4602-9a57-23009be33879_1200x600.png" width="1200" height="600" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/6102779b-f9ab-4602-9a57-23009be33879_1200x600.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:600,&quot;width&quot;:1200,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:379005,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!-e6K!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6102779b-f9ab-4602-9a57-23009be33879_1200x600.png 424w, https://substackcdn.com/image/fetch/$s_!-e6K!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6102779b-f9ab-4602-9a57-23009be33879_1200x600.png 848w, https://substackcdn.com/image/fetch/$s_!-e6K!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6102779b-f9ab-4602-9a57-23009be33879_1200x600.png 1272w, https://substackcdn.com/image/fetch/$s_!-e6K!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6102779b-f9ab-4602-9a57-23009be33879_1200x600.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.ankursnewsletter.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Subscribe today to join our community of 1819 AI enthusiasts and business leaders for more insider takes and practical advice on implementing AI in business.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h2>Happy New Year and welcome to 2025!</h2><p>This past year has been filled with remarkable advancements, and we're thrilled to embark on another year of discovering and analyzing the latest AI developments with our valued community of over 1,800 subscribers.</p><div><hr></div><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!R9ua!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!R9ua!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!R9ua!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png" width="1456" height="246" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/bea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:246,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:273651,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:&quot;&quot;,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!R9ua!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1456w" sizes="100vw" loading="lazy" fetchpriority="high"></picture><div></div></div></a></figure></div><p>Last year, I started Multimodal, a Generative AI company that helps organizations automate complex, knowledge-based workflows using AI Agents. Check it out <a href="https://www.multimodal.dev/">here</a>.</p><div><hr></div><h2>Looking back, leaping forward</h2><p>2024 was a year of unprecedented growth in AI capabilities and enterprise deployment. We've covered groundbreaking LLMs, fine-tuning techniques, comparison of revolutionary applications, and the rise of AI-driven startups reshaping industries. Here are some posts you all loved last year:</p><ol><li><p><a href="https://www.ankursnewsletter.com/p/comparative-analysis-gemma-7b-vs">Comparative Analysis: Gemma 7B vs. Mistral 7B</a></p></li><li><p><a href="https://www.ankursnewsletter.com/p/google-tpus-vs-aws-trainium-and-inferentia">Google TPUs vs. AWS Trainium &amp; Inferentia vs. NVIDIA GPUs</a></p></li><li><p><a href="https://www.ankursnewsletter.com/p/bard-vs-gemini-what-are-the-differences">Bard Vs. Gemini: What Are the Differences?</a></p></li><li><p><a href="https://www.ankursnewsletter.com/p/aws-bedrock-vs-google-vertex-ai-vs">AWS Bedrock vs. Google Vertex AI vs Azure OpenAI: A comparative overview</a></p></li><li><p><a href="https://www.ankursnewsletter.com/p/how-to-calculate-ai-roi-for-your">How to calculate AI ROI for your business</a></p></li></ol><p>Your engagement has been the cornerstone of our newsletter's success. In 2025, I'll explore new AI developments and effective techniques to break the barriers to implementing AI at scale. As we toast the new year, I&#8217;m filled with gratitude for your continued support and enthusiasm.</p><p>See you next week,</p><p>Ankur.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.ankursnewsletter.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">If you liked reading this post, consider subscribing for more AI content.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p>]]></content:encoded></item><item><title><![CDATA[Microsoft Copilot vs Salesforce Einstein vs IBM Watson vs Oracle AI: A comparison]]></title><description><![CDATA[Compare enterprise AI solutions: Microsoft Copilot, Salesforce Einstein, IBM Watson, and Oracle AI. Explore their strengths, weaknesses, and industry applications.]]></description><link>https://www.ankursnewsletter.com/p/microsoft-copilot-vs-salesforce-einstein</link><guid isPermaLink="false">https://www.ankursnewsletter.com/p/microsoft-copilot-vs-salesforce-einstein</guid><dc:creator><![CDATA[Ankur A. Patel]]></dc:creator><pubDate>Thu, 19 Dec 2024 15:09:19 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!EX_M!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd491f337-a377-40cc-aae0-4d56f81467af_1200x600.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!EX_M!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd491f337-a377-40cc-aae0-4d56f81467af_1200x600.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!EX_M!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd491f337-a377-40cc-aae0-4d56f81467af_1200x600.png 424w, https://substackcdn.com/image/fetch/$s_!EX_M!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd491f337-a377-40cc-aae0-4d56f81467af_1200x600.png 848w, https://substackcdn.com/image/fetch/$s_!EX_M!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd491f337-a377-40cc-aae0-4d56f81467af_1200x600.png 1272w, https://substackcdn.com/image/fetch/$s_!EX_M!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd491f337-a377-40cc-aae0-4d56f81467af_1200x600.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!EX_M!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd491f337-a377-40cc-aae0-4d56f81467af_1200x600.png" width="1200" height="600" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d491f337-a377-40cc-aae0-4d56f81467af_1200x600.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:600,&quot;width&quot;:1200,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:395140,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!EX_M!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd491f337-a377-40cc-aae0-4d56f81467af_1200x600.png 424w, https://substackcdn.com/image/fetch/$s_!EX_M!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd491f337-a377-40cc-aae0-4d56f81467af_1200x600.png 848w, https://substackcdn.com/image/fetch/$s_!EX_M!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd491f337-a377-40cc-aae0-4d56f81467af_1200x600.png 1272w, https://substackcdn.com/image/fetch/$s_!EX_M!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd491f337-a377-40cc-aae0-4d56f81467af_1200x600.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.ankursnewsletter.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Subscribe today to join our community of 1797 AI enthusiasts and business leaders for more insider takes and practical advice on implementing AI in business.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h2>Key Takeaways</h2><p>1. Microsoft Copilot excels in <strong>productivity enhancement</strong> and integration with the Microsoft 365 suite.</p><p>2. Salesforce Einstein dominates the <strong>CRM market</strong> delivering billions of AI-powered predictions daily across its products.</p><p>3. IBM Watson stands out for its <strong>advanced natural language processing and cognitive computing capabilities</strong>.</p><p>4. Oracle AI focuses on operational efficiency and <strong>data-driven decision-making</strong>, offering over 50 new AI agents across various business applications.</p><p>5. Each AI solution has <strong>unique strengths and weaknesses</strong>, making the choice dependent on an organization's existing infrastructure, specific needs, and long-term strategic goals.</p><div><hr></div><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!R9ua!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!R9ua!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!R9ua!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png" width="1456" height="246" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/bea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:246,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:273651,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:&quot;&quot;,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!R9ua!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!R9ua!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbea33aed-89f0-40b8-a936-63093bfba6f2_2257x382.png 1456w" sizes="100vw" loading="lazy" fetchpriority="high"></picture><div></div></div></a></figure></div><p>Last year, I started Multimodal, a Generative AI company that helps organizations automate complex, knowledge-based workflows using AI Agents. Check it out <a href="https://www.multimodal.dev/">here</a>.</p><div><hr></div><p>When building enterprise AI applications, you&#8217;ll face a choice between which AI suite to go with, especially if you&#8217;re not looking to build from scratch in-house. Depending on what applications you need, this choice can widely vary.</p><p>In this article, I will break down the strengths and weaknesses of the leading enterprise-focused AI solutions: Microsoft Copilot, Salesforce Einstein, IBM Watson, and Oracle AI.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Pnqc!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f2a9d4f-3efb-442b-b03c-1790fdacf5e3_1820x346.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Pnqc!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f2a9d4f-3efb-442b-b03c-1790fdacf5e3_1820x346.png 424w, https://substackcdn.com/image/fetch/$s_!Pnqc!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f2a9d4f-3efb-442b-b03c-1790fdacf5e3_1820x346.png 848w, https://substackcdn.com/image/fetch/$s_!Pnqc!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f2a9d4f-3efb-442b-b03c-1790fdacf5e3_1820x346.png 1272w, https://substackcdn.com/image/fetch/$s_!Pnqc!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f2a9d4f-3efb-442b-b03c-1790fdacf5e3_1820x346.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Pnqc!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f2a9d4f-3efb-442b-b03c-1790fdacf5e3_1820x346.png" width="1456" height="277" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8f2a9d4f-3efb-442b-b03c-1790fdacf5e3_1820x346.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:277,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:118889,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Pnqc!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f2a9d4f-3efb-442b-b03c-1790fdacf5e3_1820x346.png 424w, https://substackcdn.com/image/fetch/$s_!Pnqc!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f2a9d4f-3efb-442b-b03c-1790fdacf5e3_1820x346.png 848w, https://substackcdn.com/image/fetch/$s_!Pnqc!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f2a9d4f-3efb-442b-b03c-1790fdacf5e3_1820x346.png 1272w, https://substackcdn.com/image/fetch/$s_!Pnqc!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8f2a9d4f-3efb-442b-b03c-1790fdacf5e3_1820x346.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><h2>Microsoft Copilot</h2><h3>Overview of Copilot's Capabilities</h3><p><a href="https://copilot.microsoft.com/">Microsoft Copilot</a> is one of the leading AI-powered productivity tools, seamlessly integrating with the Microsoft 365 suite to revolutionize how businesses operate. <strong>At its core, Copilot leverages advanced natural language processing and generation capabilities</strong>, powered by large language models like GPT-4, to understand and respond to user queries with remarkable accuracy.</p><ul><li><p><strong>Natural Language Processing (NLP)</strong>: Copilot can interpret complex prompts and generate human-like responses, making it an invaluable asset for tasks ranging from drafting emails to analyzing financial reports.</p></li><li><p><strong>Integration with Microsoft 365 Suite</strong>: Copilot is deeply embedded in Microsoft 365 apps, including Word, Excel, PowerPoint, Outlook, and Teams, providing a unified AI assistant across the entire productivity ecosystem.</p></li></ul><ul><li><p><strong>Real-time data analysis and insights</strong>: By leveraging the Microsoft Graph, Copilot can access and analyze vast amounts of organizational data in real-time, offering insights that drive informed decision-making.</p></li></ul><h2>Salesforce Einstein</h2><p><a href="https://www.salesforce.com/artificial-intelligence/">Salesforce Einstein</a> is a powerful AI platform that revolutionizes customer relationship management (CRM) by integrating advanced artificial intelligence capabilities into Salesforce's ecosystem. As the first comprehensive AI for CRM, Einstein delivers over 80 billion AI-powered predictions daily across Salesforce products.</p><p>Einstein's predictive analytics capabilities are a game-changer for businesses seeking to stay ahead of market trends and customer behaviors:</p><ul><li><p><strong>Einstein lead scoring</strong>: Uses AI to predict which leads are most likely to convert, allowing sales teams to prioritize high-potential opportunities.</p></li><li><p><strong>Einstein opportunity insights</strong>: Analyzes customer interactions and purchase history to provide intelligent insights into sales opportunities.</p></li><li><p><strong>Einstein forecasting</strong>: Leverages AI for accurate sales forecasting, enabling better resource allocation and decision-making.</p></li></ul><p>Einstein's seamless integration with the Salesforce CRM platform enhances core CRM functionality:</p><ul><li><p><strong>Native integration</strong>: Built directly into the Salesforce platform, allowing users to leverage AI capabilities within existing workflows.</p></li><li><p><strong>Automated workflow</strong>: Streamlines CRM processes by intelligently routing tasks, alerts, and notifications based on predefined criteria.</p></li><li><p><strong>Personalization at scale</strong>: Analyzes customer data within the CRM to deliver highly personalized experiences, from tailored product recommendations to customized marketing campaigns.</p></li></ul><h2>IBM Watson</h2><h3>Watson's key capabilities</h3><p><a href="https://www.ibm.com/watson">IBM Watson</a> stands out in the AI landscape with its powerful cognitive computing capabilities, positioning it as a formidable player in the CRM market alongside giants like Salesforce and Microsoft Dynamics. At its core, Watson excels in three key areas:</p><p>1.<strong> Natural Language Processing</strong>: Watson's advanced NLP capabilities enable it to comprehend and analyze unstructured data from various sources, including customer communications, social media, and support tickets. <strong>This allows businesses to gain deeper insights into customer sentiment, preferences, and behavior.</strong></p><p>2. <strong>Machine learning and cognitive computing</strong>: Watson's machine learning algorithms continuously improve their performance by learning from data and user interactions. This adaptive capability is crucial for businesses seeking to stay ahead in the dynamic CRM market. Watson can identify patterns and trends in customer data, helping sales teams uncover new sales opportunities and optimize their strategies. </p><p>3. <strong>Advanced text analytics</strong>: Watson's text analytics capabilities go beyond simple keyword matching. It can extract meaningful insights from vast amounts of textual data, including emails, chat logs, and customer feedback. This feature is particularly valuable for service teams handling complex case management scenarios.</p><p>Watson's integration capabilities allow it to work seamlessly with various third-party apps and platforms, making it a versatile addition to existing technology stacks. While it may not have the native integration that Microsoft Dynamics CRM offers, or the extensive sales and marketing cloud features of Salesforce, Watson's AI capabilities provide a unique edge in data analysis and decision-making support.</p><p>For businesses looking to leverage AI in their CRM strategy, Watson offers a powerful set of tools that can enhance customer relationships, automate routine tasks, and provide deep, actionable insights. <strong>Its ability to process and understand natural language makes it particularly suited for companies dealing with large volumes of customer communications across multiple channels.</strong></p><h2>Oracle AI</h2><h3>Oracle AI's Core Offerings</h3><p><a href="https://www.oracle.com/artificial-intelligence/">Oracle</a> has deeply integrated AI capabilities throughout its Fusion Cloud Applications Suite, offering customers AI-driven outcomes without the need to change existing applications or interfaces. This embedded approach includes:</p><ul><li><p>Over 50 new AI agents across ERP, HCM, SCM, and CX applications.</p></li><li><p>Generative AI-assisted authoring for tasks like developing order acknowledgment emails and creating executive summaries for quotes and proposals.</p></li><li><p><strong>AI-powered smart operations workbench for real-time insights into work orders and shift reporting in manufacturing environments.</strong></p></li></ul><p>Oracle provides flexible options for organizations to develop and deploy custom AI models:</p><ul><li><p>Access to foundational models from providers like Cohere and Meta in a managed environment.</p></li><li><p>Fine-tuning capabilities to adapt models to specific business requirements.</p></li><li><p>OCI Generative AI service for seamless integration of language models into various use cases.</p></li></ul><p>Oracle AI is tightly integrated with Oracle Cloud Infrastructure (OCI), offering several advantages:</p><ul><li><p>Unified security, governance, and data management across Fusion Cloud Applications and other business-critical workloads.</p></li><li><p>High-performance, low-cost GPU cluster technology suitable for demanding scenarios like large language processing.</p></li><li><p>Fusion Applications Environment Management, an OCI native service providing end-to-end lifecycle management of Fusion Applications environments.</p></li></ul><p>Oracle's approach to AI is characterized by its focus on delivering practical business value through a fully integrated technology stack. <strong>This strategy allows organizations to leverage AI capabilities across their entire operations, from core business applications to custom-built solutions, all within a secure and scalable cloud infrastructure.</strong></p><h2>Comparative analysis</h2><p>Microsoft Copilot, Salesforce Einstein, IBM Watson, and Oracle AI each offer unique strengths and weaknesses in the AI-driven enterprise solutions landscape, particularly for financial services and insurance industries.</p><h3>Microsoft Copilot</h3><h4>Strengths</h4><p>- Seamless integration with Microsoft 365 suite, enhancing productivity across familiar tools</p><p>- Leverages large language models for advanced natural language processing</p><p>- Studies show Copilot can help employees save up to 20 hours per week on mundane tasks[1]</p><h4>Weaknesses</h4><p>- Primarily focused on productivity and may lack some industry-specific features</p><p>- Effectiveness depends on existing Microsoft ecosystem adoption</p><h3>Salesforce Einstein</h3><h4>Strengths</h4><p>- Deep integration with Salesforce CRM, holding a 19.5% market share in the CRM space.</p><p>- Robust custom AI model development capabilities.</p><p>- Delivers over 80 billion AI-powered predictions daily across Salesforce products.</p><h4>Weaknesses</h4><p>- May require significant investment in Salesforce ecosystem</p><p>- Less effective for organizations not heavily reliant on CRM functionality</p><h3>IBM Watson</h3><h4>Strengths</h4><p>- Advanced natural language processing and cognitive computing capabilities.</p><p>- Strong focus on data analysis and insights.</p><p>- Proven track record in financial services, with fraud detection accuracy rates up to 95%.</p><h4>Weaknesses</h4><p>- Can be complex to implement and may require specialized expertise.</p><p>- May not integrate as seamlessly with common office productivity tools.</p><h3>Oracle AI</h3><h4>Strengths</h4><p>- Seamless integration with Oracle Cloud Infrastructure.</p><p>- Focus on operational efficiency and data-driven decision making.</p><p>- Over 50 new AI agents across ERP, HCM, SCM, and CX applications.</p><h4>Weaknesses</h4><p>- Primarily beneficial for organizations already invested in the Oracle ecosystem.</p><p>- May have a steeper learning curve for non-Oracle users.</p><h3>Industry-specific considerations</h3><h4>Data security and compliance requirements</h4><p>Financial services and insurance industries face stringent regulatory requirements. All four platforms offer robust security features, but their approaches differ:</p><p>- Microsoft Copilot adheres to Microsoft's existing privacy, security, and compliance commitments.</p><p>- Salesforce Einstein provides enhanced security features, crucial for financial data protection.</p><p>- <strong>IBM Watson offers advanced fraud detection capabilities, essential for financial institutions.</strong></p><p>- Oracle AI focuses on unified security and governance across its cloud applications.</p><h4>Integration with legacy systems</h4><p>The financial sector often relies on legacy systems, making integration crucial</p><p>- Microsoft Copilot's integration with widely-used Microsoft products eases adoption</p><p>- Salesforce Einstein may require more significant changes to existing workflows</p><p>- IBM Watson offers flexible integration options but may need more customization</p><p>- Oracle AI provides smooth integration for organizations already using Oracle systems</p><h4>Scalability and customization options</h4><p>Each platform offers different levels of scalability and customization:</p><p>- Microsoft Copilot scales with existing Microsoft 365 subscriptions</p><p>- Salesforce Einstein allows for extensive customization of AI models</p><p>- IBM Watson offers highly scalable solutions for large-scale data analysis</p><p>- Oracle AI provides scalable solutions within the Oracle Cloud ecosystem</p><div><hr></div><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!-kh9!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!-kh9!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!-kh9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png" width="1456" height="246" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:246,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:172293,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:&quot;&quot;,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!-kh9!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 424w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 848w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 1272w, https://substackcdn.com/image/fetch/$s_!-kh9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe5bf25b3-270f-453e-90d6-c3cc122bc905_2257x382.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p>I also host an AI podcast and content series called &#8220;Pioneers.&#8221; This series takes you on an enthralling journey into the minds of AI visionaries, founders, and CEOs who are at the forefront of innovation through AI in their organizations.</p><p>To learn more, please visit <a href="https://www.multimodal.dev/blog">Pioneers</a> on Beehiiv.</p><div><hr></div><h2>Wrapping Up</h2><p>As enterprise AI solutions continue to evolve, financial services and insurance industries must adapt to leverage their potential while addressing implementation challenges. Here&#8217;s what you should keep in mind while building and deploying AI solutions:</p><p>- <strong>Emerging trends</strong>: Generative AI capabilities are set to transform customer service and risk assessment processes.</p><p>- <strong>Holistic AI strategy</strong>: Organizations need to integrate AI across their entire technology stack for seamless data flow and consistent customer experiences.</p><p>- <strong>Balancing automation and expertise</strong>: While AI can handle routine tasks, human judgment remains crucial for complex decision-making and customer trust.</p><p>I&#8217;ll return next week with more on building and deploying enterprise AI applications.</p><p>Until then,</p><p>Ankur.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.ankursnewsletter.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">If you liked reading this post, consider subscribing for more AI content.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item></channel></rss>