<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[AI Unlimited]]></title><description><![CDATA[AI Unlimited delivers concise, actionable insights on the latest advancements in artificial intelligence. Stay updated on cutting-edge research, innovative tools, and practical applications, making AI accessible for enthusiasts and professionals alike.]]></description><link>https://ai-unltd.com</link><image><url>https://substackcdn.com/image/fetch/$s_!A0L1!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8255d44-7fec-4989-82c0-961bf24ab714_256x256.png</url><title>AI Unlimited</title><link>https://ai-unltd.com</link></image><generator>Substack</generator><lastBuildDate>Sun, 19 Apr 2026 01:05:45 GMT</lastBuildDate><atom:link href="https://ai-unltd.com/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[AI Unlimited]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[aiunltd@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[aiunltd@substack.com]]></itunes:email><itunes:name><![CDATA[Arjun]]></itunes:name></itunes:owner><itunes:author><![CDATA[Arjun]]></itunes:author><googleplay:owner><![CDATA[aiunltd@substack.com]]></googleplay:owner><googleplay:email><![CDATA[aiunltd@substack.com]]></googleplay:email><googleplay:author><![CDATA[Arjun]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[Why You Should Worry About AI]]></title><description><![CDATA[Embracing Pragmatic Worry: How to Thrive in the Age of AI]]></description><link>https://ai-unltd.com/p/why-you-should-worry-about-ai</link><guid isPermaLink="false">https://ai-unltd.com/p/why-you-should-worry-about-ai</guid><dc:creator><![CDATA[Arjun]]></dc:creator><pubDate>Sat, 18 Oct 2025 19:08:09 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!YN2a!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F04695aed-7870-494b-b7e5-691448dc0e01_1536x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!YN2a!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F04695aed-7870-494b-b7e5-691448dc0e01_1536x1024.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!YN2a!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F04695aed-7870-494b-b7e5-691448dc0e01_1536x1024.png 424w, https://substackcdn.com/image/fetch/$s_!YN2a!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F04695aed-7870-494b-b7e5-691448dc0e01_1536x1024.png 848w, https://substackcdn.com/image/fetch/$s_!YN2a!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F04695aed-7870-494b-b7e5-691448dc0e01_1536x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!YN2a!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F04695aed-7870-494b-b7e5-691448dc0e01_1536x1024.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!YN2a!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F04695aed-7870-494b-b7e5-691448dc0e01_1536x1024.png" width="1456" height="971" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/04695aed-7870-494b-b7e5-691448dc0e01_1536x1024.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:971,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2734425,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://ai-unltd.com/i/176506594?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F04695aed-7870-494b-b7e5-691448dc0e01_1536x1024.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!YN2a!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F04695aed-7870-494b-b7e5-691448dc0e01_1536x1024.png 424w, https://substackcdn.com/image/fetch/$s_!YN2a!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F04695aed-7870-494b-b7e5-691448dc0e01_1536x1024.png 848w, https://substackcdn.com/image/fetch/$s_!YN2a!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F04695aed-7870-494b-b7e5-691448dc0e01_1536x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!YN2a!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F04695aed-7870-494b-b7e5-691448dc0e01_1536x1024.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Worry about AI</figcaption></figure></div><p>Artificial intelligence isn&#8217;t science fiction anymore; it&#8217;s the driving force reshaping our world in 2025. This technology is remaking our lives at a pace humanity has never experienced before. <strong>The question isn&#8217;t whether AI will change everything; it already has</strong>. The challenge is whether you&#8217;re preparing for what comes next.</p><h2>AI is Genuinely Life-Changing</h2><p>Across the world, companies are seeing gains that once felt impossible.</p><ul><li><p><strong>PGP Glass</strong>: Employees save <strong>30&#8211;40 minutes</strong> every day with Microsoft 365 Copilot.</p></li><li><p><strong>Sandvik</strong>: Productivity is up <strong>as much as 30%</strong> thanks to AI assistants in manufacturing.</p></li><li><p><strong>Toshiba</strong>: <strong>5.6 hours</strong> saved per person, per month&#8212;across <strong>10,000</strong> employees.</p></li></ul><p>These aren&#8217;t small tweaks. They&#8217;re a new way of working.</p><ul><li><p>Studies from <strong>Harvard, Wharton, and MIT</strong> show top consultants improved performance by <strong>40%</strong> using tools like ChatGPT.</p></li><li><p>Workers with AI skills now earn a <strong>56% wage premium</strong> (up from <strong>25%</strong> last year).</p></li><li><p><strong>PwC (2025)</strong>: Industries best positioned for AI saw <strong>3&#215; faster growth</strong> in revenue per employee.</p></li></ul><p><strong>Bottom line:</strong> AI isn&#8217;t just helpful, it&#8217;s a force multiplier for people and businesses.</p><h2>Job Automation is Easier Than Ever</h2><p>What makes this wave of automation different from previous technological disruptions is the sheer accessibility and sophistication of AI tools. <strong>Job automation is no longer the exclusive domain of large enterprises with massive IT budgets, it&#8217;s available to any organization with an internet connection.</strong></p><p>The barrier to entry has collapsed. Generative AI platforms like ChatGPT, Claude, and Google Gemini can be deployed with minimal technical expertise. Companies can now automate knowledge work such as writing, analysis, research, coding, design, these were the tasks that were considered safe from automation just years ago. <strong>According to research, 80% of the U.S. workforce could have at least 10% of their tasks automated very easily.</strong></p><p><strong>The timeline for major disruption has accelerated to 2027-2028, making immediate adaptation strategies essential</strong>. What once would have taken decades now unfolds in years or even months. The rapid pace of AI capability improvement means that <strong>tasks considered &#8220;too complex for AI&#8221; today may be fully automatable within 24 months</strong>.</p><h2>Worry About AI Every DAY!</h2><p>Let&#8217;s worry, but in a pragmatic way. Here&#8217;s the uncomfortable truth: worry isn&#8217;t weakness; it&#8217;s wisdom.</p><p>I propose a simple but powerful practice: <strong>Dedicate five minutes each day to worry about AI. Not anxious, paralyzing worry but a productive, action-oriented concern</strong>. Use these five minutes to ask yourself critical questions:</p><ul><li><p>Which of my current tasks could AI automate within a month?</p></li><li><p>What skills am I developing that will remain valuable in an AI-augmented workplace?</p></li><li><p>How is AI changing my industry, and what does that mean for my role?</p></li><li><p>What new opportunities is AI creating that I could capitalize on?</p></li><li><p>Am I learning how to effectively collaborate with AI tools, or am I resisting them?</p></li></ul><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://ai-unltd.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">You can also subscribe to my newsletter to hear about all new things &#8220;AI&#8221;.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>This daily practice serves three purposes:</p><ol><li><p>It keeps AI transformation at the forefront of your awareness, preventing complacency.</p></li><li><p>It creates space for strategic thinking rather than reactive scrambling when disruption arrives at your doorstep.</p></li><li><p>It transforms abstract anxiety into concrete action plans.</p></li></ol><h2>AI boosts productivity ONLY when used in the right places</h2><ul><li><p><strong>Real gains, real numbers</strong></p><ul><li><p>Typical generative-AI users save <strong>~5.4%</strong> of work time (<strong>~2.2 hours/week</strong> on a 40-hour schedule).</p></li><li><p>On the right tasks, the lift can be <strong>huge</strong>.</p></li><li><p>Dev teams using <strong>GitHub Copilot</strong> report <strong>~20%</strong> higher productivity&#8212;lower costs, faster releases.</p></li></ul></li><li><p><strong>Skills are changing fassssttttt &#128640;</strong></p><ul><li><p>By <strong>2030</strong>, employers expect <strong>39%</strong> of key job skills to change.</p></li><li><p>In AI-exposed roles, skills are shifting <strong>66% faster</strong> than in less-exposed roles (up from <strong>25%</strong> last year).</p></li><li><p>Translation: even with AI, <strong>keep upskilling</strong>.</p></li></ul></li><li><p><strong>The real multiplier</strong></p><ul><li><p>Don&#8217;t just do the same work faster&#8212;use AI to <strong>handle the operational grind</strong>.</p></li><li><p>Free people to focus on <strong>bigger, higher-impact problems</strong>.</p></li><li><p>The best results come from <strong>redesigning workflows around AI</strong>, not just adding tools.</p></li></ul></li></ul><h2>Conclusion: Channel Worry to Action</h2><p>The key to navigating this landscape lies in embracing pragmatic worry; not as a source of paralysis, but as a catalyst for empowerment. By structuring your daily worry, you convert abstract fears into a disciplined framework for growth, aligning with projections that AI could boost global GDP by 1.5% by 2035 while creating 78 million new jobs for those equipped to seize them.</p><p>AI will reshape everything but it doesn&#8217;t have to reshape you by surprise. Worry the right way, act with intent, and let AI multiply your impact.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://ai-unltd.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading AI Unlimited! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><div><hr></div><p><strong>Sources:</strong></p><ul><li><p><a href="https://budgetmodel.wharton.upenn.edu/issues/2025/9/8/projected-impact-of-generative-ai-on-future-productivity-growth">The Projected Impact of Generative AI on Future Productivity Growth</a></p></li><li><p><a href="https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5316265">AI Job Displacement Analysis (2025-2030)</a></p></li><li><p><a href="https://www.microsoft.com/en-us/microsoft-cloud/blog/2025/07/24/ai-powered-success-with-1000-stories-of-customer-transformation-and-innovation/">AI business impact: Microsoft AI use cases</a></p></li><li><p><a href="https://www.daniweb.com/community-center/op-ed/540748/ai-boosts-human-performance-by-another-40-who-will-profit">AI Boosts Human Performance by Another 40%: Who Will Profit?</a></p></li></ul>]]></content:encoded></item><item><title><![CDATA[Best Summarization Prompt: Summarizing Long Texts with AI]]></title><description><![CDATA[How to Summarize Long Texts with AI Tools Like ChatGPT]]></description><link>https://ai-unltd.com/p/best-summarization-prompt-summarizing</link><guid isPermaLink="false">https://ai-unltd.com/p/best-summarization-prompt-summarizing</guid><dc:creator><![CDATA[Arjun]]></dc:creator><pubDate>Fri, 29 Nov 2024 21:52:32 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!LjNm!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7ea750ef-2b0d-4fed-badb-bea05e24bd5a_1792x1024.webp" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!LjNm!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7ea750ef-2b0d-4fed-badb-bea05e24bd5a_1792x1024.webp" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!LjNm!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7ea750ef-2b0d-4fed-badb-bea05e24bd5a_1792x1024.webp 424w, https://substackcdn.com/image/fetch/$s_!LjNm!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7ea750ef-2b0d-4fed-badb-bea05e24bd5a_1792x1024.webp 848w, https://substackcdn.com/image/fetch/$s_!LjNm!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7ea750ef-2b0d-4fed-badb-bea05e24bd5a_1792x1024.webp 1272w, https://substackcdn.com/image/fetch/$s_!LjNm!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7ea750ef-2b0d-4fed-badb-bea05e24bd5a_1792x1024.webp 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!LjNm!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7ea750ef-2b0d-4fed-badb-bea05e24bd5a_1792x1024.webp" width="1456" height="832" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/7ea750ef-2b0d-4fed-badb-bea05e24bd5a_1792x1024.webp&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:832,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:431158,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/webp&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!LjNm!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7ea750ef-2b0d-4fed-badb-bea05e24bd5a_1792x1024.webp 424w, https://substackcdn.com/image/fetch/$s_!LjNm!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7ea750ef-2b0d-4fed-badb-bea05e24bd5a_1792x1024.webp 848w, https://substackcdn.com/image/fetch/$s_!LjNm!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7ea750ef-2b0d-4fed-badb-bea05e24bd5a_1792x1024.webp 1272w, https://substackcdn.com/image/fetch/$s_!LjNm!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7ea750ef-2b0d-4fed-badb-bea05e24bd5a_1792x1024.webp 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>In today&#8217;s fast-paced world, processing large volumes of information is a common challenge. Using efficient prompts to summarize long texts efficiently can save time, improve comprehension, and enhance productivity. In this article, we&#8217;ll explore a practical and actionable prompt to summarization using a structured prompt designed for in-depth analysis and clarity.</p><h2>The Perfect Prompt for Summarizing Long Texts</h2><p>To unlock the full potential of AI for summarization, you need the right prompt. Below is a carefully crafted example that guides AI to generate concise, insightful summaries by addressing key points in the text.</p><pre><code>&lt;text&gt;
&lt;/text&gt;

1. Analyse the above text enclosed in &lt;text&gt; tags and generate 5 essential questions that, when answered, capture the main points and core meaning of the text

2. When formulating your questions:
     a. Address the central theme or argument
     b. Identify key supporting ideas
     c. Highlight important facts or evidence 
     d. Reveal the author's purpose or perspective
     e. Explore any significant implications or conclusions

3. Answer all of your generated questions one-by-one in detail

4. Compile all the answers into a paragraph and present as a summary</code></pre><p>Add the long text you want to summarize between the &lt;text&gt;&lt;/text&gt; tags.</p><p><a href="https://chatgpt.com/share/674a35ae-202c-8008-b22e-f1e0e9931fad">Here is an example of it in action</a></p><h2>How does it work?</h2><h4>Generate Essential Questions</h4><p>The LLM begins by generating five questions based on the text. These questions address the core argument, supporting ideas, important facts, the author&#8217;s perspective, and implications. This process ensures that the key points are captured comprehensively.</p><h4>Provide Detailed Answers</h4><p>The LLM answers each question thoroughly, breaking down the text into digestible insights. This step transforms complex information into manageable chunks.</p><h4>Compile Answers into a Summary</h4><p>Finally, the answers are compiled into a cohesive paragraph that captures the essence of the text. This step ensures clarity and fluency, delivering a concise summary.</p><h2>Why This Prompt is Highly Efficient?</h2><p>It combines a systematic framework with a focus on depth and clarity. By structuring the process through essential questions and detailed answers, it mimics a human&#8217;s critical thinking and comprehension process</p><h4><strong>Comprehensive Coverage</strong></h4><p>By breaking the text into central theme, supporting ideas, evidence, purpose, and implications, the prompt ensures no key aspect is overlooked.</p><h4><strong>Adaptive to Complexity</strong></h4><p>The open-ended nature of the questions makes it versatile enough to handle texts of varying lengths and complexities, from simple articles to dense academic papers.</p><h4><strong>Encourages Depth</strong></h4><p>The detailed question-answer process avoids superficial summaries, ensuring the core meaning and intricacies of the text are captured.</p><h4><strong>Promotes Clarity and Relevance</strong></h4><p>Each generated question and answer is directly tied to the text, reducing the risk of irrelevant or vague summaries.</p><h4><strong>Logical Flow</strong></h4><p>The final paragraph synthesizes the extracted information into a cohesive and digestible form, making it ready for immediate use.</p><div><hr></div><p>Summarizing long texts is no longer a time-consuming chore, thanks to LLMs like ChatGPT. With the right prompt, you can extract meaningful insights, improve your understanding, and stay productive. Try this prompt the next time you encounter an overwhelming article or document, and experience the efficiency of AI-driven summarization firsthand.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://ai-unltd.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading AI Unlimited! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p><p></p>]]></content:encoded></item><item><title><![CDATA[Reduce GPT Costs with Prompt Compression]]></title><description><![CDATA[Prompt compression is a technique that reduces the length of inputs given to large language models (LLMs).]]></description><link>https://ai-unltd.com/p/reduce-gpt-costs-with-prompt-compression</link><guid isPermaLink="false">https://ai-unltd.com/p/reduce-gpt-costs-with-prompt-compression</guid><dc:creator><![CDATA[Arjun]]></dc:creator><pubDate>Thu, 11 Jul 2024 16:58:55 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!I0i_!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F223668c7-3bb7-4ddc-a967-c37ef85cd2ec_1792x1024.webp" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!I0i_!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F223668c7-3bb7-4ddc-a967-c37ef85cd2ec_1792x1024.webp" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!I0i_!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F223668c7-3bb7-4ddc-a967-c37ef85cd2ec_1792x1024.webp 424w, https://substackcdn.com/image/fetch/$s_!I0i_!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F223668c7-3bb7-4ddc-a967-c37ef85cd2ec_1792x1024.webp 848w, https://substackcdn.com/image/fetch/$s_!I0i_!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F223668c7-3bb7-4ddc-a967-c37ef85cd2ec_1792x1024.webp 1272w, https://substackcdn.com/image/fetch/$s_!I0i_!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F223668c7-3bb7-4ddc-a967-c37ef85cd2ec_1792x1024.webp 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!I0i_!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F223668c7-3bb7-4ddc-a967-c37ef85cd2ec_1792x1024.webp" width="1456" height="832" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/223668c7-3bb7-4ddc-a967-c37ef85cd2ec_1792x1024.webp&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:832,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:359902,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/webp&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!I0i_!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F223668c7-3bb7-4ddc-a967-c37ef85cd2ec_1792x1024.webp 424w, https://substackcdn.com/image/fetch/$s_!I0i_!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F223668c7-3bb7-4ddc-a967-c37ef85cd2ec_1792x1024.webp 848w, https://substackcdn.com/image/fetch/$s_!I0i_!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F223668c7-3bb7-4ddc-a967-c37ef85cd2ec_1792x1024.webp 1272w, https://substackcdn.com/image/fetch/$s_!I0i_!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F223668c7-3bb7-4ddc-a967-c37ef85cd2ec_1792x1024.webp 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Prompt compression is a technique that reduces the length of inputs given to large language models (LLMs). It aims to maintain output quality while using fewer tokens. This is important because LLM APIs (such as OpenAI, Anthropic, etc.) charge based on the number of tokens processed.</p><p>LLMs break down text into tokens, which are chunks of characters. Each token has a cost associated with it. Longer prompts use more tokens, leading to higher costs. Prompt compression helps you stay within token limits and reduce processing time.</p><h2>Main Compression Techniques</h2><p>There are three primary methods for compressing prompts:</p><ol><li><p>Knowledge Distillation: This involves summarization or rewriting the sentences to reduce the number of tokens and make things more precise and short.<br><strong>Example:</strong><br>Original: "Explain the process of photosynthesis in detail, including the light-dependent and light-independent reactions, and how this process is crucial for life on Earth."<br>Compressed: "Describe photosynthesis: light reactions, dark reactions, importance for life."</p></li><li><p>Encoding: This technique transforms text into a format that we might not be able to comprehend easily, but LLMs can make sense of them.<br><strong>Example:</strong><br>Original: "The quick brown fox jumps over the lazy dog. This sentence contains every letter of the English alphabet."<br>Encoded: &#8220;VGhlIHF1aWNrIGJyb3duIGZveCBqdW1wcyBvdmVyIHRoZ&#8230;&#8221;</p></li><li><p>Filtering: This method removes unnecessary parts of prompts. It keeps only the most relevant information, reducing token count. This can be done at various levels, such as sentences, phrases, or tokens.<br><strong>Example:</strong><br>Original: "Can you please provide me with a comprehensive and detailed explanation of the fundamental principles of quantum mechanics, including its historical development and key concepts such as superposition, entanglement, and wave-particle duality?"<br>Filtered: "Explain quantum mechanics: principles, history, superposition, entanglement, wave-particle duality."</p></li></ol><p>Each of these techniques can significantly reduce the number of tokens in a prompt while preserving its core meaning. The choice of technique depends on your specific use case and the type of information you're working with.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://ai-unltd.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">AI Newsletter Today is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h2>Implementing Prompt Compression</h2><p>To implement prompt compression, you can use various tools and libraries. One popular option is <a href="https://www.microsoft.com/en-us/research/blog/llmlingua-innovating-llm-efficiency-with-prompt-compression/">LLMLingua</a>, developed by Microsoft. It uses advanced techniques to refine prompts into key components.</p><p>Here are examples of how you might implement basic versions of our three main compression techniques using Python.</p><h4><strong>Knowledge Distillation (Simplified example using a pre-trained model)</strong></h4><pre><code>from transformers import pipeline

def compress_prompt_distill(prompt, max_length=50):
    summarizer = pipeline("summarization", model="facebook/bart-large-cnn")
    summary = summarizer(prompt, max_length=max_length, min_length=10, do_sample=False)
    return summary[0]['summary_text']

original_prompt = "Explain the process of photosynthesis in detail, including the light-dependent and light-independent reactions, and how this process is crucial for life on Earth."
distilled_prompt = compress_prompt_distill(original_prompt)
print(f"Distilled prompt: {distilled_prompt}")</code></pre><h4><strong>Encoding (Using sentence embeddings)</strong></h4><pre><code>from sentence_transformers import SentenceTransformer

def compress_prompt_encode(prompt):
    model = SentenceTransformer('all-MiniLM-L6-v2')
    return model.encode(prompt)

original_prompt = "The quick brown fox jumps over the lazy dog."
encoded_prompt = compress_prompt_encode(original_prompt)
print(f"Encoded prompt (first 5 values): {encoded_prompt[:5]}")</code></pre><h4><strong>Filtering (Keyword Extraction)</strong></h4><pre><code>import nltk
from nltk.corpus import stopwords
from nltk.tokenize import word_tokenize

nltk.download('punkt')
nltk.download('stopwords')

def compress_prompt_filter(prompt, num_keywords=10):
    words = word_tokenize(prompt.lower())
    stop_words = set(stopwords.words('english'))
    keywords = [word for word in words if word.isalnum() and word not in stop_words]
    freq_dist = nltk.FreqDist(keywords)
    top_keywords = [word for word, _ in freq_dist.most_common(num_keywords)]
    return " ".join(top_keywords)

original_prompt = "Prompt compression is a technique used to optimize inputs given to large language models (LLMs) by reducing their length while maintaining output quality and relevance."
compressed_prompt = compress_prompt_filter(original_prompt)
print(f"Filtered prompt: {compressed_prompt}")</code></pre><p>When implementing prompt compression, keep these best practices in mind:</p><ul><li><p>Maintain a balance between compression and preserving essential information.</p></li><li><p>Test compressed prompts to ensure they still produce high-quality outputs.</p></li><li><p>Be aware that excessive compression can lead to loss of important nuances.</p></li></ul><p>Common challenges include handling diverse types of input data and ensuring consistent output quality. Overcome these by testing your compression methods on a variety of prompts and fine-tuning your approach.</p><p>Case studies often show significant cost reductions, sometimes up to 50% or more, without major impact on output quality. However, results can vary based on the specific use case and compression method.</p><p>The field of prompt compression is rapidly evolving. New techniques are emerging, such as dynamic compression ratios and context-aware compression. In the future, we may see prompt compression integrated directly into LLM architectures, making it even more efficient and effective.</p><p>By implementing prompt compression, you can significantly reduce your LLM token costs while maintaining the quality of your AI applications. Start with simple techniques and gradually explore more advanced methods as you become comfortable with the concept.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://ai-unltd.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">AI Newsletter Today is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Tutorial: Chat with PDFs on your Google Drive]]></title><description><![CDATA[6 easy steps to chat with your document in Google Drive. It costs less than a ChatGPT subscription and is more secure.]]></description><link>https://ai-unltd.com/p/tutorial-chat-with-pdfs-in-google</link><guid isPermaLink="false">https://ai-unltd.com/p/tutorial-chat-with-pdfs-in-google</guid><dc:creator><![CDATA[Arjun]]></dc:creator><pubDate>Mon, 05 Feb 2024 19:00:16 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!crj2!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff1ce701d-c1a4-49a5-ab22-914df6dc22cc_1792x1024.webp" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!crj2!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff1ce701d-c1a4-49a5-ab22-914df6dc22cc_1792x1024.webp" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!crj2!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff1ce701d-c1a4-49a5-ab22-914df6dc22cc_1792x1024.webp 424w, https://substackcdn.com/image/fetch/$s_!crj2!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff1ce701d-c1a4-49a5-ab22-914df6dc22cc_1792x1024.webp 848w, https://substackcdn.com/image/fetch/$s_!crj2!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff1ce701d-c1a4-49a5-ab22-914df6dc22cc_1792x1024.webp 1272w, https://substackcdn.com/image/fetch/$s_!crj2!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff1ce701d-c1a4-49a5-ab22-914df6dc22cc_1792x1024.webp 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!crj2!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff1ce701d-c1a4-49a5-ab22-914df6dc22cc_1792x1024.webp" width="1456" height="832" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f1ce701d-c1a4-49a5-ab22-914df6dc22cc_1792x1024.webp&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:832,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:537200,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/webp&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!crj2!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff1ce701d-c1a4-49a5-ab22-914df6dc22cc_1792x1024.webp 424w, https://substackcdn.com/image/fetch/$s_!crj2!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff1ce701d-c1a4-49a5-ab22-914df6dc22cc_1792x1024.webp 848w, https://substackcdn.com/image/fetch/$s_!crj2!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff1ce701d-c1a4-49a5-ab22-914df6dc22cc_1792x1024.webp 1272w, https://substackcdn.com/image/fetch/$s_!crj2!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff1ce701d-c1a4-49a5-ab22-914df6dc22cc_1792x1024.webp 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Unlock the ability to interact with PDFs in your Google Drive as if you're having a conversation with them. This simple guide is crafted for anyone who wants immediate results without delving into the technicalities. Get ready to chat with your documents!</p><div class="preformatted-block" data-component-name="PreformattedTextBlockToDOM"><label class="hide-text" contenteditable="false">Text within this block will maintain its original spacing when published</label><pre class="text">Find a sample google colab document all the way at the bottom of this article.</pre></div><h2>What You'll Need</h2><p>Before we begin, ensure you have the following:</p><ul><li><p>A PDF uploaded to your Google Drive.</p></li><li><p>A Google Colab account ready for use.</p></li><li><p>An <a href="https://platform.openai.com/api-keys">OpenAI account to access your API key</a>. Using OpenAI's API comes with costs, so please <a href="https://platform.openai.com/usage">monitor your usage</a> regularly.</p></li><li><p>And, you need to subscribe to this newsletter&#8230; Not mandatory, but please? &#128517;</p></li></ul><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://ai-unltd.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://ai-unltd.com/subscribe?"><span>Subscribe now</span></a></p><h2>Setting Up Your Workspace</h2><h4>Step 1: Open Your Google Colab Notebook</h4><p>Visit <a href="https://colab.google/">Google Colab</a>, login with your google account and create a new colab notebook. That should look something like this. &#128071;</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!djbW!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7c7e4325-a426-40ec-9e0a-06d2088632cd_3024x1708.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!djbW!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7c7e4325-a426-40ec-9e0a-06d2088632cd_3024x1708.png 424w, https://substackcdn.com/image/fetch/$s_!djbW!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7c7e4325-a426-40ec-9e0a-06d2088632cd_3024x1708.png 848w, https://substackcdn.com/image/fetch/$s_!djbW!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7c7e4325-a426-40ec-9e0a-06d2088632cd_3024x1708.png 1272w, https://substackcdn.com/image/fetch/$s_!djbW!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7c7e4325-a426-40ec-9e0a-06d2088632cd_3024x1708.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!djbW!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7c7e4325-a426-40ec-9e0a-06d2088632cd_3024x1708.png" width="1456" height="822" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/7c7e4325-a426-40ec-9e0a-06d2088632cd_3024x1708.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:822,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:203875,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!djbW!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7c7e4325-a426-40ec-9e0a-06d2088632cd_3024x1708.png 424w, https://substackcdn.com/image/fetch/$s_!djbW!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7c7e4325-a426-40ec-9e0a-06d2088632cd_3024x1708.png 848w, https://substackcdn.com/image/fetch/$s_!djbW!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7c7e4325-a426-40ec-9e0a-06d2088632cd_3024x1708.png 1272w, https://substackcdn.com/image/fetch/$s_!djbW!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7c7e4325-a426-40ec-9e0a-06d2088632cd_3024x1708.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h4>Step 2: Connect to Google Drive</h4><p>Within your new notebook, look for the file icon on the left sidebar. Click on it and wait for a few seconds for the Colab environment to load.</p><p>Then click the Google Drive symbol to mount your drive. This should ask you a couple of permissions to enable you to connect with your google drive.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!lUkT!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8611ea7c-659b-4fe4-8693-77bd9c4c14f6_1012x734.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!lUkT!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8611ea7c-659b-4fe4-8693-77bd9c4c14f6_1012x734.png 424w, https://substackcdn.com/image/fetch/$s_!lUkT!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8611ea7c-659b-4fe4-8693-77bd9c4c14f6_1012x734.png 848w, https://substackcdn.com/image/fetch/$s_!lUkT!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8611ea7c-659b-4fe4-8693-77bd9c4c14f6_1012x734.png 1272w, https://substackcdn.com/image/fetch/$s_!lUkT!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8611ea7c-659b-4fe4-8693-77bd9c4c14f6_1012x734.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!lUkT!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8611ea7c-659b-4fe4-8693-77bd9c4c14f6_1012x734.png" width="1012" height="734" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8611ea7c-659b-4fe4-8693-77bd9c4c14f6_1012x734.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:734,&quot;width&quot;:1012,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:52459,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!lUkT!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8611ea7c-659b-4fe4-8693-77bd9c4c14f6_1012x734.png 424w, https://substackcdn.com/image/fetch/$s_!lUkT!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8611ea7c-659b-4fe4-8693-77bd9c4c14f6_1012x734.png 848w, https://substackcdn.com/image/fetch/$s_!lUkT!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8611ea7c-659b-4fe4-8693-77bd9c4c14f6_1012x734.png 1272w, https://substackcdn.com/image/fetch/$s_!lUkT!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8611ea7c-659b-4fe4-8693-77bd9c4c14f6_1012x734.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Once this is done, you should see two things:</p><ol><li><p>A drive folder which is nothing but your google drive that&#8217;s connected to the colab notebook.</p></li><li><p>A piece of code that can be used to connect to your colab notebook. The code should look something like below.</p></li></ol><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!kj_C!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c9cb2c1-1d66-4a2c-a900-bcc4dc9abda1_1352x734.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!kj_C!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c9cb2c1-1d66-4a2c-a900-bcc4dc9abda1_1352x734.png 424w, https://substackcdn.com/image/fetch/$s_!kj_C!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c9cb2c1-1d66-4a2c-a900-bcc4dc9abda1_1352x734.png 848w, https://substackcdn.com/image/fetch/$s_!kj_C!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c9cb2c1-1d66-4a2c-a900-bcc4dc9abda1_1352x734.png 1272w, https://substackcdn.com/image/fetch/$s_!kj_C!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c9cb2c1-1d66-4a2c-a900-bcc4dc9abda1_1352x734.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!kj_C!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c9cb2c1-1d66-4a2c-a900-bcc4dc9abda1_1352x734.png" width="1352" height="734" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/3c9cb2c1-1d66-4a2c-a900-bcc4dc9abda1_1352x734.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:734,&quot;width&quot;:1352,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:56063,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!kj_C!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c9cb2c1-1d66-4a2c-a900-bcc4dc9abda1_1352x734.png 424w, https://substackcdn.com/image/fetch/$s_!kj_C!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c9cb2c1-1d66-4a2c-a900-bcc4dc9abda1_1352x734.png 848w, https://substackcdn.com/image/fetch/$s_!kj_C!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c9cb2c1-1d66-4a2c-a900-bcc4dc9abda1_1352x734.png 1272w, https://substackcdn.com/image/fetch/$s_!kj_C!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3c9cb2c1-1d66-4a2c-a900-bcc4dc9abda1_1352x734.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Now run the code by pressing the play button on the left side of the code. This might again ask you for a few permissions, but once it finishes running, your google drive will be connected with the colab notebook.</p><h4>Step 3: Installing Embedchain package</h4><p>Before we install, lets add a new cell by hovering on the centre of the last cell and clicking on &#8220;+ Code&#8221; button.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Brdc!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Feca7b30c-60e9-4ff4-baa6-d5d3027619bd_1352x734.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Brdc!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Feca7b30c-60e9-4ff4-baa6-d5d3027619bd_1352x734.png 424w, https://substackcdn.com/image/fetch/$s_!Brdc!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Feca7b30c-60e9-4ff4-baa6-d5d3027619bd_1352x734.png 848w, https://substackcdn.com/image/fetch/$s_!Brdc!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Feca7b30c-60e9-4ff4-baa6-d5d3027619bd_1352x734.png 1272w, https://substackcdn.com/image/fetch/$s_!Brdc!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Feca7b30c-60e9-4ff4-baa6-d5d3027619bd_1352x734.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Brdc!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Feca7b30c-60e9-4ff4-baa6-d5d3027619bd_1352x734.png" width="1352" height="734" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/eca7b30c-60e9-4ff4-baa6-d5d3027619bd_1352x734.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:734,&quot;width&quot;:1352,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:82334,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Brdc!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Feca7b30c-60e9-4ff4-baa6-d5d3027619bd_1352x734.png 424w, https://substackcdn.com/image/fetch/$s_!Brdc!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Feca7b30c-60e9-4ff4-baa6-d5d3027619bd_1352x734.png 848w, https://substackcdn.com/image/fetch/$s_!Brdc!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Feca7b30c-60e9-4ff4-baa6-d5d3027619bd_1352x734.png 1272w, https://substackcdn.com/image/fetch/$s_!Brdc!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Feca7b30c-60e9-4ff4-baa6-d5d3027619bd_1352x734.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption"></figcaption></figure></div><p>Once we do that, let&#8217;s add the following code to the new cell and run it again by clicking the play button.</p><pre><code>!pip install embedchain</code></pre><p>This will generate a lot of output messages, wait for it to finish, scroll all the way to the bottom and look for the "Successfully installed" message to confirm everything went smoothly.</p><h4>Step 4: Getting Ready to Chat</h4><p>Import Packages and Load Your OpenAI API Key</p><p>Here&#8217;s where you bring in the tools you need. Add the following code to a new cell:</p><pre><code>import os
from embedchain import App

os.environ["OPENAI_API_KEY"] = "sk-yourapikey"
bot = App()</code></pre><p>Don't forget to replace <code>"sk-yourapikey"</code> with the actual key you get from your <a href="https://platform.openai.com/api-keys">OpenAI account</a>.</p><h4>Step 5: Adding Your PDF to the Conversation</h4><p>Add Your PDF to the Bot</p><p>With this step, you tell the bot about the PDF you want to chat with. Add the following code to a new cell and run it:</p><pre><code>bot.add('/content/drive/MyDrive/PDFs/grammy_awards.pdf', data_type='pdf_file')</code></pre><p>I have my PDF (grammy_awards.pdf) in a folder called &#8220;PDFs&#8221; in my google drive. Make sure to update the above code to match the path of your pdf in your google drive.</p><h4>Step 6: Let&#8217;s start the chat!</h4><p>Start Chatting with Your PDF!</p><p>It's showtime! Ask your PDF anything by adding the below code to a new cell and running it:</p><pre><code>bot.query('Who hosted the grammy awards of the year 2014?')</code></pre><p>Replace the above question with your own and you can start chatting with the PDF in your google drive!</p><p>Here&#8217;s a sample output:</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!UMTU!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ab04a7a-92dc-4cd4-b172-a1b0b357ec34_1602x334.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!UMTU!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ab04a7a-92dc-4cd4-b172-a1b0b357ec34_1602x334.png 424w, https://substackcdn.com/image/fetch/$s_!UMTU!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ab04a7a-92dc-4cd4-b172-a1b0b357ec34_1602x334.png 848w, https://substackcdn.com/image/fetch/$s_!UMTU!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ab04a7a-92dc-4cd4-b172-a1b0b357ec34_1602x334.png 1272w, https://substackcdn.com/image/fetch/$s_!UMTU!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ab04a7a-92dc-4cd4-b172-a1b0b357ec34_1602x334.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!UMTU!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ab04a7a-92dc-4cd4-b172-a1b0b357ec34_1602x334.png" width="1456" height="304" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/3ab04a7a-92dc-4cd4-b172-a1b0b357ec34_1602x334.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:304,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:55141,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!UMTU!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ab04a7a-92dc-4cd4-b172-a1b0b357ec34_1602x334.png 424w, https://substackcdn.com/image/fetch/$s_!UMTU!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ab04a7a-92dc-4cd4-b172-a1b0b357ec34_1602x334.png 848w, https://substackcdn.com/image/fetch/$s_!UMTU!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ab04a7a-92dc-4cd4-b172-a1b0b357ec34_1602x334.png 1272w, https://substackcdn.com/image/fetch/$s_!UMTU!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3ab04a7a-92dc-4cd4-b172-a1b0b357ec34_1602x334.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p><strong><a href="https://colab.research.google.com/drive/1h6W_tf47ebtXH8yfFXojVgVwp1qxjAY7?usp=sharing">Here is the link to a sample code on Google colab.</a></strong></p><h2>Conclusion</h2><p>You've just stepped into the future of document interaction. With these few steps, you can effortlessly extract information from your PDFs. Remember to monitor your OpenAI API usage to manage any associated costs.</p><p>Enjoy your new-found efficiency in data retrieval and happy chatting with your documents!</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://ai-unltd.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">AI Newsletter Today is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Rephrase and Respond (RaR): A New Way to Prompt ChatGPT for Accurate Responses]]></title><description><![CDATA[RaR methodology surpasses Chain-of-Thought in multiple benchmarks.]]></description><link>https://ai-unltd.com/p/rephrase-and-response-rar-a-new-way</link><guid isPermaLink="false">https://ai-unltd.com/p/rephrase-and-response-rar-a-new-way</guid><dc:creator><![CDATA[Arjun]]></dc:creator><pubDate>Sun, 17 Dec 2023 19:58:28 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!X2Qi!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F309026d1-4a70-4252-831e-e17469ed2c9e_1792x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!X2Qi!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F309026d1-4a70-4252-831e-e17469ed2c9e_1792x1024.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!X2Qi!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F309026d1-4a70-4252-831e-e17469ed2c9e_1792x1024.png 424w, https://substackcdn.com/image/fetch/$s_!X2Qi!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F309026d1-4a70-4252-831e-e17469ed2c9e_1792x1024.png 848w, https://substackcdn.com/image/fetch/$s_!X2Qi!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F309026d1-4a70-4252-831e-e17469ed2c9e_1792x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!X2Qi!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F309026d1-4a70-4252-831e-e17469ed2c9e_1792x1024.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!X2Qi!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F309026d1-4a70-4252-831e-e17469ed2c9e_1792x1024.png" width="1456" height="832" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/309026d1-4a70-4252-831e-e17469ed2c9e_1792x1024.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:832,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1164141,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!X2Qi!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F309026d1-4a70-4252-831e-e17469ed2c9e_1792x1024.png 424w, https://substackcdn.com/image/fetch/$s_!X2Qi!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F309026d1-4a70-4252-831e-e17469ed2c9e_1792x1024.png 848w, https://substackcdn.com/image/fetch/$s_!X2Qi!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F309026d1-4a70-4252-831e-e17469ed2c9e_1792x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!X2Qi!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F309026d1-4a70-4252-831e-e17469ed2c9e_1792x1024.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h3>Contents</h3><ol><li><p>Introduction</p></li><li><p>Understanding the Need for Better Questioning in LLMs</p></li><li><p>The RaR Method Explained <em>(With prompt examples)</em></p></li><li><p>Benefits of RaR in Enhancing LLM Responses</p></li><li><p>RaR vs. Chain-of-Thought (CoT) Method</p></li><li><p>Conclusion</p></li></ol><h2>Introduction</h2><p>In the evolving landscape of artificial intelligence, particularly with Large Language Models (LLMs) like ChatGPT, <strong>the method of interaction is crucial in eliciting accurate and meaningful responses</strong>. The Rephrase and Respond (RaR) method enhances the way we prompt these models by offering a new perspective on optimizing our dialogue with AI. By rephrasing questions and responses, <em>RaR aims to address common communication gaps, ensuring that LLMs like ChatGPT understand and reply with greater precision</em>. <strong>The best part is, that this method has been tested and benchmarked on GPT-4, one of the top LLMs out there.</strong></p><h2>Understanding the Need for Better Questioning in LLMs</h2><blockquote><p>The effectiveness of ChatGPT heavily relies on the quality of the prompts it receives. </p></blockquote><p><em>When posed with the query, &#8220;Was Mother Teresa born on an even month?&#8221; GPT-4 might mistakenly assert that August is an odd month.</em></p><p>Often, users face challenges in framing questions that elicit accurate and comprehensive responses. This gap arises due to the inherent limitations of LLMs in understanding nuanced or complex queries. <strong>Misinterpretations, lack of context, and ambiguous phrasing lead to suboptimal responses</strong>.</p><p>Rephrase and Respond (RaR) method emerges as a solution, aimed at refining the way questions are posed to LLMs. <strong>By focusing on the art of rephrasing, RaR addresses the core issue of communication breakdown, ensuring that the queries align more closely with the LLM's processing capabilities.</strong></p><h2>The RaR Method Explained <em>(With prompt examples)</em></h2><p>The RaR method elevates the effectiveness of LLMs like ChatGPT. <strong>It encompasses two distinct strategies: the one-step RaR and the two-step RaR.</strong></p><p><strong>One-Step RaR</strong>: <em>This involves the LLM rephrasing the user's query into a single, more precise question before responding</em>. This rephrasing aims to clarify the query's intent and ensure a correct understanding, leading to a more accurate answer.</p><pre><code>"{question}"
Rephrase and expand the question, and respond.</code></pre><p><a href="https://chat.openai.com/share/0ab795a3-d640-4e20-b841-89b45b5b306b">One-Step RaR ChatGPT Link</a></p><p><strong>Two-Step RaR</strong>: In this approach, the <em>LLM first rephrases the query and then, in a separate step, responds to this refined query</em>. This two-step process allows for even greater clarity and specificity in understanding and addressing the user's needs.</p><pre><code># Step 1

"{question}"
Given the above question, rephrase and expand it to help you
do better answering. Maintain all information in the original question. 

# Step 2

(original) "{question}"
(rephrased) "{rephrased_question}"
Use your answer for the rephrased question to answer the original question.</code></pre><p><a href="https://chat.openai.com/share/87fc6da5-d233-4342-83bb-b9927d9f8d96">Two-Step RaR ChatGPT Link</a></p><p><strong>The key advantage of RaR over traditional questioning methods is its focus on question clarity and precision</strong>.</p><p><em>With RaR, the LLM first seeks to disentangle the query's ambiguities, asking for clarifications or rephrasing it to grasp the user's true intent. This ensures that the response is more relevant and informative.</em></p><h2>Benefits of RaR in Enhancing LLM Responses</h2><blockquote><p>The RaR method significantly elevates the performance of ChatGPT.</p></blockquote><p>Key benefits include:</p><ol><li><p><strong>Improved Accuracy</strong>: RaR, in both its one-step and two-step forms, <em>substantially increases response accuracy in ChatGPT</em>. This is evident in the improved performance metrics across a range of tasks.</p></li><li><p><strong>Enhanced Relevance</strong>: Responses become more contextually relevant, as ChatGPT better grasps the specifics of the query through rephrasing.</p></li><li><p><strong>Increased Efficiency</strong>: RaR can streamline interactions, <em>particularly in cases where initial queries may be too vague or broad</em>.</p></li><li><p><strong>Versatility in Application</strong>: <em>RaR's adaptability makes it suitable for diverse applications, from simple Q&amp;A to complex problem-solving scenarios.</em></p></li></ol><p><a href="https://arxiv.org/abs/2311.04205">Research paper link.</a></p><h2>RaR vs. Chain-of-Thought (CoT) Method</h2><p>Comparing the RaR method with the Chain-of-Thought (CoT) approach reveals distinct features and applications:</p><ol><li><p><strong>Approach</strong>:</p><ul><li><p>RaR focuses on refining queries through <strong>rephrasing and iterative clarification</strong>, enhancing understanding before responding.</p></li><li><p>CoT involves the <strong>LLM explicating its reasoning process step-by-step</strong>, akin to a human solving a problem out loud.</p></li></ul></li><li><p><strong>Accuracy and Efficiency</strong>:</p><ul><li><p>RaR aims to <strong>increase accuracy by ensuring the question is well-understood before answering</strong>, which can be more efficient in obtaining precise information.</p></li><li><p>CoT, by <strong>elaborating the thought process, may provide deeper insights into complex problems</strong> but can be more time-consuming.</p></li></ul></li><li><p><strong>User Interaction</strong>:</p><ul><li><p>RaR <strong>encourages active user participation in refining the query</strong>, making it more interactive.</p></li><li><p>CoT is more AI-centric, with the <strong>model displaying its reasoning without direct user intervention</strong> in the thought process.</p></li></ul></li><li><p><strong>Applicability</strong>:</p><ul><li><p>RaR is <strong>versatile, and suitable for a wide range of queries</strong>, especially where precision and clarity are key.</p></li><li><p>CoT <strong>excels in scenarios requiring detailed explanations or step-by-step reasoning</strong>, such as complex problem-solving.</p></li></ul></li></ol><p>Each method has its strengths, and choosing between them depends on the specific needs of the interaction, whether it's <strong>clarity and precision (RaR) or detailed understanding and explanation (CoT)</strong>.</p><p><em>Additionally, CoT and RaR can be combined to get the best of both.</em></p><h2>Conclusion</h2><p>RaR enables LLMs to rephrase queries for better comprehension and more accurate responses. As we move forward, <strong>RaR promises to revolutionize LLM interactions, making them more intuitive, efficient, and aligned with human communication</strong>. We encourage you to experiment with RaR in their interactions with ChatGPT, to experience first-hand the advancements it brings to human-AI communication.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://ai-unltd.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">AI Newsletter Today is a reader-supported publication. To read more, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p><p></p><p></p>]]></content:encoded></item><item><title><![CDATA[The Rise of Mistral AI]]></title><description><![CDATA[Mistral AI stands as a testament to the power of open-source models]]></description><link>https://ai-unltd.com/p/the-rise-of-mistral-ai</link><guid isPermaLink="false">https://ai-unltd.com/p/the-rise-of-mistral-ai</guid><dc:creator><![CDATA[Arjun]]></dc:creator><pubDate>Thu, 14 Dec 2023 09:00:52 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!U-Ku!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc5444da8-8e98-4fb8-be4c-83be2eb5fcb3_1792x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!U-Ku!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc5444da8-8e98-4fb8-be4c-83be2eb5fcb3_1792x1024.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!U-Ku!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc5444da8-8e98-4fb8-be4c-83be2eb5fcb3_1792x1024.png 424w, https://substackcdn.com/image/fetch/$s_!U-Ku!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc5444da8-8e98-4fb8-be4c-83be2eb5fcb3_1792x1024.png 848w, https://substackcdn.com/image/fetch/$s_!U-Ku!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc5444da8-8e98-4fb8-be4c-83be2eb5fcb3_1792x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!U-Ku!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc5444da8-8e98-4fb8-be4c-83be2eb5fcb3_1792x1024.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!U-Ku!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc5444da8-8e98-4fb8-be4c-83be2eb5fcb3_1792x1024.png" width="1456" height="832" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c5444da8-8e98-4fb8-be4c-83be2eb5fcb3_1792x1024.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:832,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:3478511,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!U-Ku!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc5444da8-8e98-4fb8-be4c-83be2eb5fcb3_1792x1024.png 424w, https://substackcdn.com/image/fetch/$s_!U-Ku!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc5444da8-8e98-4fb8-be4c-83be2eb5fcb3_1792x1024.png 848w, https://substackcdn.com/image/fetch/$s_!U-Ku!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc5444da8-8e98-4fb8-be4c-83be2eb5fcb3_1792x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!U-Ku!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc5444da8-8e98-4fb8-be4c-83be2eb5fcb3_1792x1024.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Mistral AI, a Paris-based innovator, is swiftly redefining the artificial intelligence landscape. Unlike its contemporaries, Mistral AI brings a unique blend of scientific acumen and open-source enthusiasm to the forefront. Its emergence is not just a story of technological innovation but also one of breaking barriers in a domain traditionally dominated by a few tech giants.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!QPIY!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdf306820-a61f-48c8-84aa-a83598fe1320_500x341.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!QPIY!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdf306820-a61f-48c8-84aa-a83598fe1320_500x341.png 424w, https://substackcdn.com/image/fetch/$s_!QPIY!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdf306820-a61f-48c8-84aa-a83598fe1320_500x341.png 848w, https://substackcdn.com/image/fetch/$s_!QPIY!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdf306820-a61f-48c8-84aa-a83598fe1320_500x341.png 1272w, https://substackcdn.com/image/fetch/$s_!QPIY!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdf306820-a61f-48c8-84aa-a83598fe1320_500x341.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!QPIY!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdf306820-a61f-48c8-84aa-a83598fe1320_500x341.png" width="500" height="341" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/df306820-a61f-48c8-84aa-a83598fe1320_500x341.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:341,&quot;width&quot;:500,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;Image&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Image" title="Image" srcset="https://substackcdn.com/image/fetch/$s_!QPIY!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdf306820-a61f-48c8-84aa-a83598fe1320_500x341.png 424w, https://substackcdn.com/image/fetch/$s_!QPIY!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdf306820-a61f-48c8-84aa-a83598fe1320_500x341.png 848w, https://substackcdn.com/image/fetch/$s_!QPIY!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdf306820-a61f-48c8-84aa-a83598fe1320_500x341.png 1272w, https://substackcdn.com/image/fetch/$s_!QPIY!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdf306820-a61f-48c8-84aa-a83598fe1320_500x341.png 1456w" sizes="100vw"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Source: Ramsri Goutham Golla on Twitter</figcaption></figure></div><h2>Founding Vision, Innovative AI Models, and Open-Source Philosophy</h2><p>Mistral AI (founded at the start of 2023) was established with a clear vision to democratize AI through cutting-edge models and an open-source framework. The company's dedication to transparency is evidenced by its <a href="https://mistral.ai/news/mixtral-of-experts/">Mixtral 8x7B model</a>, a high-quality sparse-mixture-of-experts (SMoE) model with an Apache 2.0 License. This model outperforms Meta&#8217;s LLAMA 2 70B model on most benchmarks with 6x faster inference.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!p_tu!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F56a5a061-c310-4ec2-a507-97f56a2061a1_1094x768.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!p_tu!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F56a5a061-c310-4ec2-a507-97f56a2061a1_1094x768.png 424w, https://substackcdn.com/image/fetch/$s_!p_tu!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F56a5a061-c310-4ec2-a507-97f56a2061a1_1094x768.png 848w, https://substackcdn.com/image/fetch/$s_!p_tu!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F56a5a061-c310-4ec2-a507-97f56a2061a1_1094x768.png 1272w, https://substackcdn.com/image/fetch/$s_!p_tu!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F56a5a061-c310-4ec2-a507-97f56a2061a1_1094x768.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!p_tu!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F56a5a061-c310-4ec2-a507-97f56a2061a1_1094x768.png" width="476" height="334.15722120658137" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/56a5a061-c310-4ec2-a507-97f56a2061a1_1094x768.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:768,&quot;width&quot;:1094,&quot;resizeWidth&quot;:476,&quot;bytes&quot;:168022,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!p_tu!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F56a5a061-c310-4ec2-a507-97f56a2061a1_1094x768.png 424w, https://substackcdn.com/image/fetch/$s_!p_tu!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F56a5a061-c310-4ec2-a507-97f56a2061a1_1094x768.png 848w, https://substackcdn.com/image/fetch/$s_!p_tu!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F56a5a061-c310-4ec2-a507-97f56a2061a1_1094x768.png 1272w, https://substackcdn.com/image/fetch/$s_!p_tu!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F56a5a061-c310-4ec2-a507-97f56a2061a1_1094x768.png 1456w" sizes="100vw"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Expanding on the open-source philosophy, Mistral AI has adopted an approach that shares not just the model but also the specifics of its architecture such as the layer dimensions and weights, as seen in their recent open-weights release. This level of detail reflects their commitment to open-source development, allowing for broad adaptability and innovation within the AI community.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!2G_e!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fffa9775c-6114-4945-a076-7170c3c8d87e_1244x668.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!2G_e!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fffa9775c-6114-4945-a076-7170c3c8d87e_1244x668.png 424w, https://substackcdn.com/image/fetch/$s_!2G_e!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fffa9775c-6114-4945-a076-7170c3c8d87e_1244x668.png 848w, https://substackcdn.com/image/fetch/$s_!2G_e!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fffa9775c-6114-4945-a076-7170c3c8d87e_1244x668.png 1272w, https://substackcdn.com/image/fetch/$s_!2G_e!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fffa9775c-6114-4945-a076-7170c3c8d87e_1244x668.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!2G_e!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fffa9775c-6114-4945-a076-7170c3c8d87e_1244x668.png" width="504" height="270.636655948553" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/ffa9775c-6114-4945-a076-7170c3c8d87e_1244x668.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:668,&quot;width&quot;:1244,&quot;resizeWidth&quot;:504,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;Image&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Image" title="Image" srcset="https://substackcdn.com/image/fetch/$s_!2G_e!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fffa9775c-6114-4945-a076-7170c3c8d87e_1244x668.png 424w, https://substackcdn.com/image/fetch/$s_!2G_e!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fffa9775c-6114-4945-a076-7170c3c8d87e_1244x668.png 848w, https://substackcdn.com/image/fetch/$s_!2G_e!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fffa9775c-6114-4945-a076-7170c3c8d87e_1244x668.png 1272w, https://substackcdn.com/image/fetch/$s_!2G_e!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fffa9775c-6114-4945-a076-7170c3c8d87e_1244x668.png 1456w" sizes="100vw"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Source: Andrej Karpathy</figcaption></figure></div><p>Not just that, Mistral has taken a novel approach to distributing their open-source models by using torrent links. This method facilitates easier and faster sharing of large files among users, which is especially beneficial for the sizable models in AI.</p><h2>Funding Success and Market Impact</h2><p>Mistral AI's journey, <a href="https://a16z.com/announcement/investing-in-mistral/">marked by a record $415 million recent funding round</a>, reflects the market's readiness for an open-source AI revolution. This funding signifies investors' faith in Mistral AI's potential to disrupt the AI industry. The substantial investment underscores the growing importance of AI technologies that are accessible, versatile, and more importantly, open for communal development.</p><h2>Challenging the Big Tech Dominance</h2><p>In an industry where large corporations often overshadow innovation, Mistral AI stands as a testament to the power of open-source models. By offering scalable and adaptable AI solutions, Mistral AI provides an alternative path to the often closed ecosystems of big tech companies, encouraging a more inclusive and diversified AI development landscape.</p><p>Mistral AI looks to challenge the likes of Meta and OpenAI with its open-source models.</p><h2>Community and Collaboration</h2><p>At the heart of Mistral AI's philosophy is the belief in the power of community-driven development (Github, Discord, etc.). By actively engaging with a global community of developers, researchers, and AI enthusiasts, Mistral AI not only enhances its models but also fosters a culture of shared knowledge and ethical AI practices. This approach has the potential to accelerate innovation and ensure AI technology evolves in a way that benefits a broader spectrum of society.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://github.com/mistralai" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!LBef!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F596e4626-b809-4217-8cbd-02509fba1c37_684x464.png 424w, https://substackcdn.com/image/fetch/$s_!LBef!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F596e4626-b809-4217-8cbd-02509fba1c37_684x464.png 848w, https://substackcdn.com/image/fetch/$s_!LBef!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F596e4626-b809-4217-8cbd-02509fba1c37_684x464.png 1272w, https://substackcdn.com/image/fetch/$s_!LBef!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F596e4626-b809-4217-8cbd-02509fba1c37_684x464.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!LBef!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F596e4626-b809-4217-8cbd-02509fba1c37_684x464.png" width="458" height="310.6900584795322" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/596e4626-b809-4217-8cbd-02509fba1c37_684x464.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:464,&quot;width&quot;:684,&quot;resizeWidth&quot;:458,&quot;bytes&quot;:42623,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:&quot;https://github.com/mistralai&quot;,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!LBef!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F596e4626-b809-4217-8cbd-02509fba1c37_684x464.png 424w, https://substackcdn.com/image/fetch/$s_!LBef!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F596e4626-b809-4217-8cbd-02509fba1c37_684x464.png 848w, https://substackcdn.com/image/fetch/$s_!LBef!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F596e4626-b809-4217-8cbd-02509fba1c37_684x464.png 1272w, https://substackcdn.com/image/fetch/$s_!LBef!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F596e4626-b809-4217-8cbd-02509fba1c37_684x464.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h2>Future Directions and Industry Predictions</h2><p>As Mistral AI forges ahead, it stands on the precipice of initiating a broad transformation within the AI industry. Their pioneering work suggests a shift towards an ecosystem where open-source AI is standard, fostering greater innovation and ethical practices.</p><p>As businesses increasingly adopt AI, Mistral's open-source models are predicted to drive down costs and spur ethical AI solutions. Soon, Mistral AI's models may well become a cornerstone in AI research, with their open-source nature accelerating the pace of AI advancements and applications in real-world scenarios.</p><h2>Conclusion: Mistral AI's Position in the Tech Ecosystem</h2><p>As Mistral AI continues to grow, it not only cements its place in the AI sector but also champions a new era of AI development. Its focus on open-source models and community collaboration sets a precedent for future AI innovations, making Mistral AI a crucial player in shaping the trajectory of AI technology.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://ai-unltd.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading the AI Newsletter Today! Subscribe for free to read more articles.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p>]]></content:encoded></item><item><title><![CDATA[Everything about the EU's AI Act]]></title><description><![CDATA[Historic Step for AI Regulation in Europe]]></description><link>https://ai-unltd.com/p/everything-about-the-eus-ai-act</link><guid isPermaLink="false">https://ai-unltd.com/p/everything-about-the-eus-ai-act</guid><dc:creator><![CDATA[Arjun]]></dc:creator><pubDate>Wed, 13 Dec 2023 10:07:44 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!k3UM!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42498e47-422b-4968-9f76-9ac870c1aaa8_1792x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!k3UM!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42498e47-422b-4968-9f76-9ac870c1aaa8_1792x1024.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!k3UM!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42498e47-422b-4968-9f76-9ac870c1aaa8_1792x1024.png 424w, https://substackcdn.com/image/fetch/$s_!k3UM!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42498e47-422b-4968-9f76-9ac870c1aaa8_1792x1024.png 848w, https://substackcdn.com/image/fetch/$s_!k3UM!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42498e47-422b-4968-9f76-9ac870c1aaa8_1792x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!k3UM!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42498e47-422b-4968-9f76-9ac870c1aaa8_1792x1024.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!k3UM!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42498e47-422b-4968-9f76-9ac870c1aaa8_1792x1024.png" width="1456" height="832" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/42498e47-422b-4968-9f76-9ac870c1aaa8_1792x1024.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:832,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2350380,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!k3UM!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42498e47-422b-4968-9f76-9ac870c1aaa8_1792x1024.png 424w, https://substackcdn.com/image/fetch/$s_!k3UM!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42498e47-422b-4968-9f76-9ac870c1aaa8_1792x1024.png 848w, https://substackcdn.com/image/fetch/$s_!k3UM!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42498e47-422b-4968-9f76-9ac870c1aaa8_1792x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!k3UM!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F42498e47-422b-4968-9f76-9ac870c1aaa8_1792x1024.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h2>Historic Step for AI Regulation in Europe</h2><p>The European Union marked a significant milestone in digital regulation with the provisional agreement on the Artificial Intelligence Act (AI Act). This groundbreaking legislation is poised to establish a robust framework ensuring the safety, legality, trustworthiness, and respect for fundamental rights within AI systems.</p><h2>Key Features of the AI Act</h2><ol><li><p><strong>Risk-Based Approach:</strong> The Act categorizes AI risks into four levels, from minimal to unacceptable, each subject to tailored rules. Systems posing unacceptable risks will be banned, barring limited exceptions.</p></li><li><p><strong>Banned and Restricted AI Applications:</strong> The Act prohibits AI systems that:</p><ul><li><p>Use biometric categorization based on sensitive characteristics (race, political beliefs, etc.).</p></li><li><p>Engage in untargeted scraping of facial images for recognition databases.</p></li><li><p>Manipulate human behaviour, exploit vulnerabilities, or enable social scoring.</p></li><li><p>Recognize emotions in workplaces and educational institutions.</p></li></ul></li><li><p><strong>Safeguards for Law Enforcement Use:</strong> Strict safeguards and narrow exceptions are stipulated for using biometric identification systems by law enforcement, including conditions like prior judicial authorization and targeted searches for serious crimes.</p></li><li><p><strong>Obligations for High-Risk AI Systems:</strong> High-risk systems, impacting health, safety, fundamental rights, etc., must undergo a fundamental rights impact assessment, among other stringent requirements.</p></li><li><p><strong>Innovation and SME Support:</strong> The Act promotes regulatory sandboxes and real-world testing to aid SMEs in developing AI solutions, free from undue pressure from industry giants.</p></li><li><p><strong>Sanctions for Non-Compliance:</strong> Companies may face substantial fines ranging from 35 million euros or 7% of global turnover to 7.5 million euros or 1.5% for non-compliance, depending on the infringement and size of the company.</p></li></ol><h2>Global Impact and Next Steps</h2><p>The EU AI Act sets a precedent akin to the General Data Protection Regulation (GDPR), aiming to shape the global AI landscape ethically and safely. The next phase involves formal adoption by the European Parliament and Council.</p><h2>Author&#8217;s Comments</h2><ul><li><p><strong>Balanced Approach:</strong> The EU's AI Act embodies a risk-aware yet experimental strategy, balancing safety with innovation.</p></li><li><p><strong>Broad Scope:</strong> It addresses a wide array of sectors, from education and workplaces to SMEs and industries.</p></li><li><p><strong>Empowering SMEs:</strong> The Act introduces regulatory sandboxes and real-world testing to mitigate the dominance of large tech companies, fostering a more equitable AI development landscape.</p></li><li><p><strong>Indirect Impact on LLMs:</strong> While not directly addressing Large Language Models (LLMs) like OpenAI's GPT or Meta's LLAMA, the Act subtly controls them through stringent data usage regulations, ensuring ethical AI development from the ground up.</p></li></ul><p>Source: <a href="https://www.europarl.europa.eu/news/en/press-room/20231206IPR15699/artificial-intelligence-act-deal-on-comprehensive-rules-for-trustworthy-ai">Artificial Intelligence Act: deal on comprehensive rules for trustworthy AI</a></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://ai-unltd.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading the AI Newsletter Today! Subscribe for free to receive AI News in your inbox.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Google's Gemini Compared with GPT-4]]></title><description><![CDATA[Google's Gemini is a remarkable leap forward in AI technology, setting new standards in various fields.]]></description><link>https://ai-unltd.com/p/googles-gemini-compared-with-gpt</link><guid isPermaLink="false">https://ai-unltd.com/p/googles-gemini-compared-with-gpt</guid><dc:creator><![CDATA[Arjun]]></dc:creator><pubDate>Fri, 08 Dec 2023 09:00:49 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!wM5_!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb1042c95-97ae-4782-931d-b8fcdff4d3cf_1018x1020.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!wM5_!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb1042c95-97ae-4782-931d-b8fcdff4d3cf_1018x1020.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!wM5_!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb1042c95-97ae-4782-931d-b8fcdff4d3cf_1018x1020.png 424w, https://substackcdn.com/image/fetch/$s_!wM5_!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb1042c95-97ae-4782-931d-b8fcdff4d3cf_1018x1020.png 848w, https://substackcdn.com/image/fetch/$s_!wM5_!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb1042c95-97ae-4782-931d-b8fcdff4d3cf_1018x1020.png 1272w, https://substackcdn.com/image/fetch/$s_!wM5_!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb1042c95-97ae-4782-931d-b8fcdff4d3cf_1018x1020.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!wM5_!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb1042c95-97ae-4782-931d-b8fcdff4d3cf_1018x1020.png" width="1018" height="1020" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b1042c95-97ae-4782-931d-b8fcdff4d3cf_1018x1020.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1020,&quot;width&quot;:1018,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1308621,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!wM5_!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb1042c95-97ae-4782-931d-b8fcdff4d3cf_1018x1020.png 424w, https://substackcdn.com/image/fetch/$s_!wM5_!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb1042c95-97ae-4782-931d-b8fcdff4d3cf_1018x1020.png 848w, https://substackcdn.com/image/fetch/$s_!wM5_!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb1042c95-97ae-4782-931d-b8fcdff4d3cf_1018x1020.png 1272w, https://substackcdn.com/image/fetch/$s_!wM5_!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb1042c95-97ae-4782-931d-b8fcdff4d3cf_1018x1020.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Google DeepMind announced the launch of Gemini on 6th December 2023, a state-of-the-art AI model set to redefine the boundaries of artificial intelligence. This latest innovation aims to significantly benefit humanity, marking a leap in how we interact with AI.</p><h2>The Genesis and Multimodal Mastery of Gemini</h2><p>Gemini is the culmination of collaborative efforts across various teams at Google, including Google Research. Designed to emulate human understanding, Gemini stands out as more than just a smart AI model; it's an intuitive and useful assistant. Its distinct edge lies in its multimodal nature, capable of comprehending and processing diverse types of information, such as text, code, audio, images, and video. This versatility allows Gemini to function efficiently across different platforms, from expansive data centres to compact mobile devices, serving a broad range of developer and enterprise applications.</p><h2>Introducing Gemini's Variants: Ultra, Pro, and Nano</h2><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!2A6P!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffbdf65db-74e8-4248-9473-dc0e6968359a_600x308.gif" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!2A6P!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffbdf65db-74e8-4248-9473-dc0e6968359a_600x308.gif 424w, https://substackcdn.com/image/fetch/$s_!2A6P!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffbdf65db-74e8-4248-9473-dc0e6968359a_600x308.gif 848w, https://substackcdn.com/image/fetch/$s_!2A6P!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffbdf65db-74e8-4248-9473-dc0e6968359a_600x308.gif 1272w, https://substackcdn.com/image/fetch/$s_!2A6P!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffbdf65db-74e8-4248-9473-dc0e6968359a_600x308.gif 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!2A6P!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffbdf65db-74e8-4248-9473-dc0e6968359a_600x308.gif" width="600" height="308" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/fbdf65db-74e8-4248-9473-dc0e6968359a_600x308.gif&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:308,&quot;width&quot;:600,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:754709,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/gif&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!2A6P!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffbdf65db-74e8-4248-9473-dc0e6968359a_600x308.gif 424w, https://substackcdn.com/image/fetch/$s_!2A6P!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffbdf65db-74e8-4248-9473-dc0e6968359a_600x308.gif 848w, https://substackcdn.com/image/fetch/$s_!2A6P!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffbdf65db-74e8-4248-9473-dc0e6968359a_600x308.gif 1272w, https://substackcdn.com/image/fetch/$s_!2A6P!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ffbdf65db-74e8-4248-9473-dc0e6968359a_600x308.gif 1456w" sizes="100vw"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Gemini 1.0 debuts in three distinct sizes:</p><ul><li><p>Gemini Ultra, the largest model, tackles highly complex tasks.</p></li><li><p>Gemini Pro offers a balanced approach for a range of tasks.</p></li><li><p>Gemini Nano, the most efficient, is perfect for on-device applications.</p></li></ul><h2>Benchmarking Excellence: Gemini Ultra vs. The Field</h2><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!VfEg!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48effde0-d550-4c19-b409-19dfbd522ac2_1968x970.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!VfEg!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48effde0-d550-4c19-b409-19dfbd522ac2_1968x970.png 424w, https://substackcdn.com/image/fetch/$s_!VfEg!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48effde0-d550-4c19-b409-19dfbd522ac2_1968x970.png 848w, https://substackcdn.com/image/fetch/$s_!VfEg!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48effde0-d550-4c19-b409-19dfbd522ac2_1968x970.png 1272w, https://substackcdn.com/image/fetch/$s_!VfEg!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48effde0-d550-4c19-b409-19dfbd522ac2_1968x970.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!VfEg!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48effde0-d550-4c19-b409-19dfbd522ac2_1968x970.png" width="1456" height="718" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/48effde0-d550-4c19-b409-19dfbd522ac2_1968x970.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:718,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:218819,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!VfEg!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48effde0-d550-4c19-b409-19dfbd522ac2_1968x970.png 424w, https://substackcdn.com/image/fetch/$s_!VfEg!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48effde0-d550-4c19-b409-19dfbd522ac2_1968x970.png 848w, https://substackcdn.com/image/fetch/$s_!VfEg!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48effde0-d550-4c19-b409-19dfbd522ac2_1968x970.png 1272w, https://substackcdn.com/image/fetch/$s_!VfEg!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F48effde0-d550-4c19-b409-19dfbd522ac2_1968x970.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Gemini Ultra, the premier model in the Gemini series, showcases exceptional performance, outshining competitors in 30 out of 32 large language model (LLM) research metrics. In the MMLU (massive multitask language understanding) test, Gemini achieves an astounding 90.04% accuracy, surpassing both human experts and other leading AI models, including GPT-4. This level of accuracy, particularly in such a broad range of subjects, from mathematics to medicine, highlights Gemini Ultra's superior understanding and problem-solving capabilities.</p><p>In specific areas like mathematics, Gemini Ultra excels with a 94.4% accuracy in the GSM8K grade-school math benchmark and 53.2% in the MATH benchmark for more advanced competition problems. These results significantly surpass GPT-4, illustrating Gemini Ultra's advanced reasoning in complex mathematical contexts. Furthermore, in coding benchmarks like HumanEval and Natural2Code for Python code generation, Gemini Ultra again sets new records, outperforming GPT-4 and showcasing its proficiency across various programming languages. These achievements demonstrate Gemini Ultra's edge in combining deep understanding with practical application, setting a new benchmark in AI technology.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!11_i!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F51eb384c-29e6-4083-b9a1-251ea2a89ddc_1600x1162.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!11_i!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F51eb384c-29e6-4083-b9a1-251ea2a89ddc_1600x1162.png 424w, https://substackcdn.com/image/fetch/$s_!11_i!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F51eb384c-29e6-4083-b9a1-251ea2a89ddc_1600x1162.png 848w, https://substackcdn.com/image/fetch/$s_!11_i!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F51eb384c-29e6-4083-b9a1-251ea2a89ddc_1600x1162.png 1272w, https://substackcdn.com/image/fetch/$s_!11_i!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F51eb384c-29e6-4083-b9a1-251ea2a89ddc_1600x1162.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!11_i!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F51eb384c-29e6-4083-b9a1-251ea2a89ddc_1600x1162.png" width="1456" height="1057" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/51eb384c-29e6-4083-b9a1-251ea2a89ddc_1600x1162.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1057,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:245389,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!11_i!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F51eb384c-29e6-4083-b9a1-251ea2a89ddc_1600x1162.png 424w, https://substackcdn.com/image/fetch/$s_!11_i!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F51eb384c-29e6-4083-b9a1-251ea2a89ddc_1600x1162.png 848w, https://substackcdn.com/image/fetch/$s_!11_i!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F51eb384c-29e6-4083-b9a1-251ea2a89ddc_1600x1162.png 1272w, https://substackcdn.com/image/fetch/$s_!11_i!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F51eb384c-29e6-4083-b9a1-251ea2a89ddc_1600x1162.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!mbbF!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F05d6629c-de2e-4da7-b1cb-907353f1c945_1354x1286.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!mbbF!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F05d6629c-de2e-4da7-b1cb-907353f1c945_1354x1286.png 424w, https://substackcdn.com/image/fetch/$s_!mbbF!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F05d6629c-de2e-4da7-b1cb-907353f1c945_1354x1286.png 848w, https://substackcdn.com/image/fetch/$s_!mbbF!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F05d6629c-de2e-4da7-b1cb-907353f1c945_1354x1286.png 1272w, https://substackcdn.com/image/fetch/$s_!mbbF!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F05d6629c-de2e-4da7-b1cb-907353f1c945_1354x1286.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!mbbF!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F05d6629c-de2e-4da7-b1cb-907353f1c945_1354x1286.png" width="1354" height="1286" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/05d6629c-de2e-4da7-b1cb-907353f1c945_1354x1286.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1286,&quot;width&quot;:1354,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:234750,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!mbbF!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F05d6629c-de2e-4da7-b1cb-907353f1c945_1354x1286.png 424w, https://substackcdn.com/image/fetch/$s_!mbbF!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F05d6629c-de2e-4da7-b1cb-907353f1c945_1354x1286.png 848w, https://substackcdn.com/image/fetch/$s_!mbbF!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F05d6629c-de2e-4da7-b1cb-907353f1c945_1354x1286.png 1272w, https://substackcdn.com/image/fetch/$s_!mbbF!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F05d6629c-de2e-4da7-b1cb-907353f1c945_1354x1286.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h2>Revolutionizing Learning and Setting New Industry Standards: The Impact of Gemini</h2><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!MSFd!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcca756f1-3fa5-457f-97fb-f166d1357dbc_1600x1138.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!MSFd!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcca756f1-3fa5-457f-97fb-f166d1357dbc_1600x1138.png 424w, https://substackcdn.com/image/fetch/$s_!MSFd!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcca756f1-3fa5-457f-97fb-f166d1357dbc_1600x1138.png 848w, https://substackcdn.com/image/fetch/$s_!MSFd!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcca756f1-3fa5-457f-97fb-f166d1357dbc_1600x1138.png 1272w, https://substackcdn.com/image/fetch/$s_!MSFd!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcca756f1-3fa5-457f-97fb-f166d1357dbc_1600x1138.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!MSFd!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcca756f1-3fa5-457f-97fb-f166d1357dbc_1600x1138.png" width="1456" height="1036" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/cca756f1-3fa5-457f-97fb-f166d1357dbc_1600x1138.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1036,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:535911,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!MSFd!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcca756f1-3fa5-457f-97fb-f166d1357dbc_1600x1138.png 424w, https://substackcdn.com/image/fetch/$s_!MSFd!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcca756f1-3fa5-457f-97fb-f166d1357dbc_1600x1138.png 848w, https://substackcdn.com/image/fetch/$s_!MSFd!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcca756f1-3fa5-457f-97fb-f166d1357dbc_1600x1138.png 1272w, https://substackcdn.com/image/fetch/$s_!MSFd!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fcca756f1-3fa5-457f-97fb-f166d1357dbc_1600x1138.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Gemini's application in educational contexts is a testament to its extraordinary capabilities. It skillfully analyzes and solves complex problems, such as interpreting and converting messy handwriting in physics problems into precise mathematical typesetting. This showcases not just its proficiency in understanding educational material, but also its potential to significantly enhance teaching and learning experiences. Gemini's ability to identify and correct errors in students' reasoning, providing accurate, step-by-step solutions, positions it as a transformative tool in education, poised to change the dynamics of learning and instruction.</p><p>Beyond the realm of education, Gemini sets new benchmarks across various industries, owing to its comprehensive training across multiple modalities like text, image, audio, and video. In text-based tasks, Gemini Pro rivals the capabilities of advanced models like GPT-4. Meanwhile, Gemini Ultra excels in areas requiring deep reasoning, reading comprehension, and expertise in STEM and coding, outperforming current standards. These achievements underscore Gemini's versatility and effectiveness, highlighting its potential to revolutionize problem-solving and task execution in a wide array of sectors. The advent of Gemini signifies not just an advancement in AI technology, but a leap forward in how AI can be applied to improve various aspects of work and education.</p><h2>Redefining Coding and Competitive Programming</h2><p>Gemini Ultra excels in understanding, explaining, and generating high-quality code across major programming languages, outperforming not only its contemporaries but also setting new benchmarks. In the realm of competitive programming, Gemini powers advanced systems like AlphaCode, which have become formidable contenders, solving complex mathematics and theoretical computer science problems with unmatched efficiency. AlphaCode 2, the upgraded version utilizing Gemini, significantly outperforms its predecessor, promising to surpass the capabilities of a substantial portion of human participants. This symbiotic fusion of human expertise and AI efficiency represents the future of coding, promising to accelerate innovation and problem-solving in the tech industry and beyond, fundamentally changing the way we approach software development and complex problem-solving challenges.</p><h2>Gemini's Integration and Accessibility</h2><p>Set to integrate with Bard, Gemini will also be accessible to developers via Google AI Studio and Google Cloud Vertex AI from December 13th, 2023. This move promises to open new frontiers in the application and development of AI technologies.</p><p><a href="https://deepmind.google/technologies/gemini/#introduction">Link to announcement.</a></p><p><a href="https://storage.googleapis.com/deepmind-media/gemini/gemini_1_report.pdf">Link to the technical report.</a></p><h2>Conclusion</h2><p>Gemini's arrival marks a transformative moment for LLMs. Its versatile applications, from education to industry benchmarks, showcase its potential. By revolutionizing education and excelling in coding tasks, Gemini augments human capabilities and drives innovation. This dynamic synergy between humans and AI heralds a future of accelerated development and groundbreaking advancements across diverse industries, all thanks to the transformative power of Gemini.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://ai-unltd.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading the AI Newsletter Today! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Beyond Chain-of-Thought: The Evolution of AI Problem-Solving with Least-to-Most Prompting]]></title><description><![CDATA[Least-to-most prompting outperforms chain-of-thought in complex problem-solving.]]></description><link>https://ai-unltd.com/p/beyond-chain-of-thought-the-evolution</link><guid isPermaLink="false">https://ai-unltd.com/p/beyond-chain-of-thought-the-evolution</guid><dc:creator><![CDATA[Arjun]]></dc:creator><pubDate>Thu, 07 Dec 2023 20:43:41 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!dgPa!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc8845cb1-77cd-4260-87f0-baa55805cf5f_2048x2048.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!dgPa!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc8845cb1-77cd-4260-87f0-baa55805cf5f_2048x2048.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!dgPa!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc8845cb1-77cd-4260-87f0-baa55805cf5f_2048x2048.png 424w, https://substackcdn.com/image/fetch/$s_!dgPa!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc8845cb1-77cd-4260-87f0-baa55805cf5f_2048x2048.png 848w, https://substackcdn.com/image/fetch/$s_!dgPa!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc8845cb1-77cd-4260-87f0-baa55805cf5f_2048x2048.png 1272w, https://substackcdn.com/image/fetch/$s_!dgPa!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc8845cb1-77cd-4260-87f0-baa55805cf5f_2048x2048.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!dgPa!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc8845cb1-77cd-4260-87f0-baa55805cf5f_2048x2048.png" width="554" height="554" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c8845cb1-77cd-4260-87f0-baa55805cf5f_2048x2048.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1456,&quot;width&quot;:1456,&quot;resizeWidth&quot;:554,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!dgPa!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc8845cb1-77cd-4260-87f0-baa55805cf5f_2048x2048.png 424w, https://substackcdn.com/image/fetch/$s_!dgPa!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc8845cb1-77cd-4260-87f0-baa55805cf5f_2048x2048.png 848w, https://substackcdn.com/image/fetch/$s_!dgPa!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc8845cb1-77cd-4260-87f0-baa55805cf5f_2048x2048.png 1272w, https://substackcdn.com/image/fetch/$s_!dgPa!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc8845cb1-77cd-4260-87f0-baa55805cf5f_2048x2048.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><strong>TL;DR</strong>: Least-to-most prompting applied to GPT-3 code-davinci-002, surpasses chain-of-thought methods in solving complex problems. This technique uses problem decomposition and subproblem solving, achieving 99% accuracy on the SCAN benchmark. This innovative approach demonstrates a leap towards deep learning systems with capabilities akin to human-like reasoning, bridging a significant gap in AI problem-solving.</p><div><hr></div><p>In the evolving world of AI prompting, the ability to solve complex problems has been a perennial challenge. Traditional chain-of-thought prompting, while effective in various reasoning tasks, often falters when faced with complexities beyond its programmed examples. This limitation sets the stage for a groundbreaking solution: least-to-most prompting.</p><h2>Bridging the Human-AI Gap</h2><p>Deep learning and human intelligence have always been worlds apart. Humans learn and reason from minimal examples, a feat AI has struggled to replicate. Chain-of-thought prompting made strides in this direction, offering improved interpretability and performance, yet it still stumbled in generalizing to more complex tasks. This is where least-to-most prompting shines, aligning more closely with how humans tackle difficult problems.</p><h2>The Rise of Least-to-Most Prompting</h2><p>Designed to tackle the challenge of easy-to-hard generalization, least-to-most prompting is a novel strategy that dissects complex problems into simpler subproblems. These subproblems are then solved sequentially, leveraging the answers of preceding ones, without necessitating extra training. It mirrors a teaching technique from educational psychology, sequentially guiding learners to build upon simpler concepts.</p><h2>Methodology and Execution</h2><p>The process involves two stages: <strong>decomposition</strong> and <strong>subproblem solving</strong>. Initially, the problem is broken down. Then, using a blend of constant examples and previously solved subquestions, each subproblem is tackled in turn. The culmination of this approach is the resolution of the original complex problem, an elegant demonstration of sequential reasoning.</p><h2>Experimental Results and Comparative Analysis</h2><p>In practice, least-to-most prompting has shown remarkable results. When applied to the GPT-3 code-davinci-002 model, it reached a staggering <strong>99% accuracy</strong> on the SCAN benchmark, significantly outperforming chain-of-thought's 16% accuracy. This impressive feat was achieved with just a handful of examples, highlighting the efficiency and effectiveness of this method.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!pQkn!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fae896a7f-6dfe-44ee-80f0-91c09f7a7308_1286x218.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!pQkn!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fae896a7f-6dfe-44ee-80f0-91c09f7a7308_1286x218.png 424w, https://substackcdn.com/image/fetch/$s_!pQkn!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fae896a7f-6dfe-44ee-80f0-91c09f7a7308_1286x218.png 848w, https://substackcdn.com/image/fetch/$s_!pQkn!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fae896a7f-6dfe-44ee-80f0-91c09f7a7308_1286x218.png 1272w, https://substackcdn.com/image/fetch/$s_!pQkn!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fae896a7f-6dfe-44ee-80f0-91c09f7a7308_1286x218.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!pQkn!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fae896a7f-6dfe-44ee-80f0-91c09f7a7308_1286x218.png" width="1286" height="218" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/ae896a7f-6dfe-44ee-80f0-91c09f7a7308_1286x218.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:218,&quot;width&quot;:1286,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:42985,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!pQkn!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fae896a7f-6dfe-44ee-80f0-91c09f7a7308_1286x218.png 424w, https://substackcdn.com/image/fetch/$s_!pQkn!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fae896a7f-6dfe-44ee-80f0-91c09f7a7308_1286x218.png 848w, https://substackcdn.com/image/fetch/$s_!pQkn!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fae896a7f-6dfe-44ee-80f0-91c09f7a7308_1286x218.png 1272w, https://substackcdn.com/image/fetch/$s_!pQkn!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fae896a7f-6dfe-44ee-80f0-91c09f7a7308_1286x218.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Accuracy of Chain-of-Thought vs Least-to-Most</figcaption></figure></div><p><a href="https://arxiv.org/abs/2205.10625">Link to paper</a></p><h2>Advantages and Limitations</h2><p>This strategy not only excels in longer and more intricate problem sets but also demonstrates integration potential with other prompting methods. However, its brilliance has its bounds. The technique&#8217;s accuracy can be marred by minor errors like concatenation mistakes, and its application is less effective in tasks requiring domain-specific decomposition.</p><h2>Conclusion</h2><p>In conclusion, least-to-most prompting is not just a step forward in AI's problem-solving capabilities; it's a leap towards bridging the gap between machine learning and the nuanced reasoning of the human mind. While it's not the ultimate solution for teaching reasoning to AI, it marks a significant advancement, nudging AI towards more efficient and human-like learning processes.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://ai-unltd.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading the AI Newsletter Today! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Get Over Chain-of-Thought, Analogical Prompting is Here! [Prompt Examples Included]]]></title><description><![CDATA[Analogical Prompting, utilising analogical reasoning, offers a refined approach.]]></description><link>https://ai-unltd.com/p/get-over-chain-of-thought-analogical</link><guid isPermaLink="false">https://ai-unltd.com/p/get-over-chain-of-thought-analogical</guid><dc:creator><![CDATA[Arjun]]></dc:creator><pubDate>Mon, 30 Oct 2023 21:01:38 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!WLAn!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8c845478-5e89-42f0-9a09-6f29d6bad214_1024x768.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!WLAn!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8c845478-5e89-42f0-9a09-6f29d6bad214_1024x768.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!WLAn!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8c845478-5e89-42f0-9a09-6f29d6bad214_1024x768.jpeg 424w, https://substackcdn.com/image/fetch/$s_!WLAn!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8c845478-5e89-42f0-9a09-6f29d6bad214_1024x768.jpeg 848w, https://substackcdn.com/image/fetch/$s_!WLAn!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8c845478-5e89-42f0-9a09-6f29d6bad214_1024x768.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!WLAn!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8c845478-5e89-42f0-9a09-6f29d6bad214_1024x768.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!WLAn!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8c845478-5e89-42f0-9a09-6f29d6bad214_1024x768.jpeg" width="592" height="444" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8c845478-5e89-42f0-9a09-6f29d6bad214_1024x768.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:768,&quot;width&quot;:1024,&quot;resizeWidth&quot;:592,&quot;bytes&quot;:455217,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!WLAn!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8c845478-5e89-42f0-9a09-6f29d6bad214_1024x768.jpeg 424w, https://substackcdn.com/image/fetch/$s_!WLAn!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8c845478-5e89-42f0-9a09-6f29d6bad214_1024x768.jpeg 848w, https://substackcdn.com/image/fetch/$s_!WLAn!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8c845478-5e89-42f0-9a09-6f29d6bad214_1024x768.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!WLAn!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8c845478-5e89-42f0-9a09-6f29d6bad214_1024x768.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><strong>TL;DR</strong>: Large Language Models (LLMs) using Chain-of-Thought (CoT) face challenges in 0-shot and few-shot reasoning. Analogical Prompting, utilizing analogical reasoning, offers a refined approach. Tapping into past experiences and generating high-level takeaways, it focuses on core problem-solving concepts. Testing on LLMs, including GPT-3.5-turbo and GPT-4, showed superior results in mathematical reasoning and code generation. As Analogical Prompting integrates knowledge generation with example crafting, it has the potential to set new benchmarks in the LLM domain, making reasoning efficient and intuitive.</p><h2>Business Implications</h2><ul><li><p>Its universal problem-solving approach can enhance product versatility, catering to a wider array of business challenges.</p></li><li><p>Enhanced AI capabilities via Analogical Prompting can optimize ROI by reducing manual efforts and expanding AI application areas.</p></li></ul><p><a href="https://chat.openai.com/share/544a97d2-e0ec-4805-97ae-71247c8cbe4a">Prompt Example 1</a></p><p><a href="https://chat.openai.com/share/9ca26c5a-5621-4c52-8f22-1d7c2c4ae1df">Prompt Example 2</a></p><div><hr></div><p>In the rapidly evolving world of artificial intelligence, Large Language Models (LLMs) have been at the forefront of pushing the boundaries of what machines can achieve. Among the strategies employed to amplify their potential, Chain-of-Thought (CoT) emerged as a frontrunner. However, like every innovation, CoT has its challenges. </p><p>Enter Analogical Prompting, a technique that promises to revolutionize the way we understand and guide LLMs.</p><h2>Chain-of-Thought (CoT) Unpacked</h2><p>At its core, CoT for LLMs showcased remarkable performance across an array of reasoning tasks. Yet, its efficacy came at a cost: the <strong>need for labelled examples to depict the reasoning process</strong>. In practice, this meant feeding LLMs with predefined instances. The 0-shot CoT offered a general instruction akin to "think step by step", serving as a generic reasoning guide. However, <strong>its broad approach sometimes fell short for intricate tasks</strong>. In contrast, few-shot CoT provided a more targeted approach, with multiple examples illustrating the reasoning process, but it had its own baggage &#8211; the onus of obtaining these labelled examples for every task. Such challenges paved the way for a quest: Could there be a technique that married the best of both worlds without their drawbacks?</p><h2>Enter Analogical Prompting: The New Kid on the Block</h2><p>What if there was a way to guide the reasoning process of LLMs seamlessly, drawing inspiration from the cognitive processes of humans? This is where Analogical Prompting enters the frame. Rooted in the principle of analogical reasoning, this method helps LLMs tap into past experiences, much like how we humans recall related problems and their solutions when faced with new challenges. For instance, calculating the area of a square becomes intuitive when one remembers the need to find its side length. By imitating this human-like reasoning, Analogical Prompting paves the way for LLMs to tackle novel problems with unprecedented efficiency.</p><h2>Advantages of Analogical Prompting Over CoT</h2><p>The magic of Analogical Prompting lies in its adaptability and self-sufficiency. By self-generating examples, it eliminates the labour of manually crafting reasoning instances for every task, a hurdle that both 0-shot and few-shot CoT grappled with. What&#8217;s more, these examples aren't just generic placeholders; they're tailored to individual problems, be it geometry or probability, ensuring a more nuanced approach. And the cherry on top? There&#8217;s no more scouring through external data sources to find relevant examples.</p><h2>Digging Deeper: How Analogical Prompting Works</h2><p>The brilliance of Analogical Prompting is not just in its outcomes but also in its methodology. Recognizing the importance of diverse examples, LLMs are directed to generate a range of 3 to 5 distinct instances in one go. But here&#8217;s the game-changer: before diving into examples, LLMs are instructed to generate high-level takeaways to complement the examples. By prioritizing this, LLMs hone in on the core concepts of the problem, ensuring that generated examples resonate with fundamental problem-solving strategies over mere superficial resemblances.</p><h2>Real-World Evaluations: Putting Analogical Prompting to the Test</h2><p>To determine its prowess, Analogical Prompting was put through its paces across a spectrum of tasks. <strong>From elementary math word conundrums and advanced high school math challenges to code generation involving intricate algorithms, and other reasoning tasks spanning logical deduction and formal fallacies</strong>. Two formidable LLMs, <strong>GPT-3.5-turbo and GPT-4</strong>, were the chosen ones for these experiments.</p><h2>Impressive Results: A Comparative Analysis</h2><p>The outcomes? Nothing short of impressive. In mathematical reasoning, Analogical Prompting left both 0-shot and few-shot CoT in the dust. It showcased similar superiority in code generation and other reasoning tasks, making a compelling case for its effectiveness. Not to mention, the fusion of knowledge generation with example crafting added another dimension to its efficacy.</p><p><a href="https://arxiv.org/abs/2310.01714">Paper Link</a></p><h2>Conclusion</h2><p>The world of LLMs stands at a fascinating juncture. While CoT laid the groundwork, Analogical Prompting is setting new benchmarks. By addressing the challenges that plagued its predecessor, it offers a glimpse into the future of machine reasoning &#8212; a future that's efficient, adaptable, and, most importantly, intuitive. As we march ahead, the exciting journey of LLMs and their capabilities is something to watch out for, and Analogical Prompting will undoubtedly be at its heart.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://ai-unltd.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading the AI Newsletter Today! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Retrieval-Augmented-Generation (RAG) vs Long-context LLMs: Which to Choose?]]></title><description><![CDATA[Is RAG superior to longer context LLMs?]]></description><link>https://ai-unltd.com/p/retrieval-augmented-generation-rag</link><guid isPermaLink="false">https://ai-unltd.com/p/retrieval-augmented-generation-rag</guid><dc:creator><![CDATA[Arjun]]></dc:creator><pubDate>Tue, 10 Oct 2023 04:30:05 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!E33R!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2bd757ea-0555-454b-b829-5080435c272d_1024x1024.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!E33R!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2bd757ea-0555-454b-b829-5080435c272d_1024x1024.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!E33R!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2bd757ea-0555-454b-b829-5080435c272d_1024x1024.jpeg 424w, https://substackcdn.com/image/fetch/$s_!E33R!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2bd757ea-0555-454b-b829-5080435c272d_1024x1024.jpeg 848w, https://substackcdn.com/image/fetch/$s_!E33R!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2bd757ea-0555-454b-b829-5080435c272d_1024x1024.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!E33R!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2bd757ea-0555-454b-b829-5080435c272d_1024x1024.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!E33R!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2bd757ea-0555-454b-b829-5080435c272d_1024x1024.jpeg" width="1024" height="1024" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/2bd757ea-0555-454b-b829-5080435c272d_1024x1024.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1024,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:181428,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!E33R!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2bd757ea-0555-454b-b829-5080435c272d_1024x1024.jpeg 424w, https://substackcdn.com/image/fetch/$s_!E33R!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2bd757ea-0555-454b-b829-5080435c272d_1024x1024.jpeg 848w, https://substackcdn.com/image/fetch/$s_!E33R!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2bd757ea-0555-454b-b829-5080435c272d_1024x1024.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!E33R!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2bd757ea-0555-454b-b829-5080435c272d_1024x1024.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><strong>TL;DR</strong>: LLMs like GPT-4, <a href="https://www.ainewsletter.today/p/extending-llama-to-32k-tokens-catching">LLAMA2 LONG</a> and Claude 2 boast Long context windows due to technological advancements in GPUs and memory-efficient exact attention. Retrieval Augmentation employs retrievers like Dragon, Contriever, and OpenAI&#8217;s text-embedding-ada-002 to provide LLMs with crucial context. A pivotal finding is that a 4K LLM augmented by retrieval can rival the efficacy of a 16K context LLM. Models like Nemo GPT-43B and LLAMA2-70B were pivotal in these comparisons, indicating that RAG could enhance the performance of an LLM irrespective of its context length.</p><h3>Business Implications</h3><ol><li><p>Retrieval-augmented LLMs can achieve top-tier AI capabilities without escalating operational costs, ensuring a competitive advantage.</p></li><li><p>Harnessing open-sourced retriever tools can drive superior AI performance, elevating customer satisfaction and trust, while keeping the costs low.</p></li><li><p>Retrieval-augmented models can provide high precision in responses, derived from expansive or latest content.</p></li></ol><div class="captioned-button-wrap" data-attrs="{&quot;url&quot;:&quot;https://ai-unltd.com/p/retrieval-augmented-generation-rag?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="CaptionedButtonToDOM"><div class="preamble"><p class="cta-caption">Like these quick takeaways? Share this article with people who might find it interesting&#8230;</p></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://ai-unltd.com/p/retrieval-augmented-generation-rag?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://ai-unltd.com/p/retrieval-augmented-generation-rag?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p></div><div><hr></div><p>In the rapidly evolving world of Large Language Models(LLMs), the two competing strategies have been contending for the spotlight: extending the context window of Large Language Models (LLMs) and augmenting LLMs with retrieval mechanisms. As the latter has been a familiar solution for a long time, it prompts us to ask two key questions:</p><ol><li><p>Is retrieval augmentation superior to longer context LLMs?</p></li><li><p>Could a synthesis of both strategies unlock unprecedented capabilities?</p></li></ol><h2>Long Context LLMs</h2><p>Long-context LLMs have recently become the focal point of discussions within research, production, and open-source domains. Spearheading this momentum is the advent of faster GPUs complemented by memory-efficient exact attention. These technologies have allowed for building Long Context LLMs with up to 32K (<a href="https://www.ainewsletter.today/p/extending-llama-to-32k-tokens-catching">LLAMA2 LONG</a>, GPT-4) and 100K &#128561; (Claude 2).</p><h2>Conceptual Understanding of Retrieval Augmentation</h2><p>Beyond the context window expansion lies the retrieval method, a well-established, alternative approach. In this method, L<strong>LMs are provided with only the relevant context fetched by a retriever</strong>, offering higher scalability and speed. This strategy essentially transforms the retrieval-augmented decoder-only LLM into a model with sparse attention, where the attention pattern is determined not beforehand but by the retriever's discernment. In simpler words, <strong>the non-retrieved data (context) is treated as irrelevant and has zero-values attention weights</strong>.</p><h2>Experiment Overview</h2><p>The primary objective is to juxtapose the retrieval augmented method (using a retriever) and the everything-in-context method (adding content to the LLM context). Venturing beyond smaller models, the investigation centres on models larger than 40B parameters.</p><p>The two models used are:</p><ul><li><p><strong>Nemo GPT-43B</strong>: A model boasting 43 billion parameters and trained on a staggering 1.1T tokens. Its training data contains a 70% English corpus, enriched by multilingual and code data sources like Common Crawl, Wikipedia, and StackExchange.</p></li><li><p><strong><a href="https://www.ainewsletter.today/p/extending-llama-to-32k-tokens-catching">LLAMA2-70B</a></strong>: A publicly accessible model with 70B parameters, trained on approximately 2T tokens dominated by English data.</p></li></ul><p>The experimental landscape spanned seven datasets, encompassing single-document QA, multi-document QA, and query-based summarization for zero-shot evaluations.</p><h2>Retrieval Mechanism Explored</h2><p>Three retrievers were employed:</p><ul><li><p><strong>Dragon</strong>: Renowned for benchmark-setting performances across supervised and zero-shot information retrieval.</p></li><li><p><strong>Contriever model</strong>: Another proficient retriever.</p></li><li><p><strong>OpenAI embedding</strong>: Specifically text-embedding-ada-002.</p></li></ul><p>The retrieval process: <strong>Questions and documents are encoded into vectors, rankings are established via methods like cosine similarity, and then the top tokens, based on their similarities, are selected and dispatched alongside the question in the LLM prompt.</strong></p><h2>Key Findings and Observations</h2><p>The findings are not either/or but an and:</p><ul><li><p>A retrieval-augmented LLM with a 4K context window astonishingly mirrored the prowess of an LLM with a 16K context window on long-context tasks, albeit with significantly reduced computational demands.</p></li><li><p>Retrieval-augmented <a href="https://www.ainewsletter.today/p/extending-llama-to-32k-tokens-catching">LLAMA2-70B with a 32K context window</a> overshadowed GPT-3.5-Turbo-16K and Davinci003 across various long-context tasks.</p></li><li><p><strong>The experiments cemented the understanding that irrespective of context window size, retrieval enhances LLM performance.</strong></p></li><li><p>Interestingly, when presented with identical evidence chunks, long-context LLMs (16K, 32K) outshine their 4K context counterparts.</p></li><li><p><strong>Public retrievers often performed better than proprietary solutions like OpenAI embeddings.</strong></p></li></ul><p><a href="https://arxiv.org/abs/2310.03025">Paper Link</a></p><h2>When To Use RAG?</h2><ul><li><p>When you have question-answering tasks based on long-form documents (PDFs, Text files).</p></li><li><p>When you have to support an LLM by passing extra information that is not already available in its context to reply to a user.</p></li><li><p>When you want the LLM to always have the latest information.</p></li><li><p>When you want the LLM to not hallucinate and make up random information &#8212; tho using a RAG is not a guarantee of eliminating hallucination completely.</p></li></ul><h2>When Is LLM Context Just Enough?</h2><ul><li><p>When answering questions based on short texts that can fit well inside the context of LLMs.</p></li><li><p>Summarization tasks that involve the LLM being able to see the whole document.</p></li><li><p>When the conversation with LLM is based on its already trained/fine-tuned data.</p></li></ul><h2>Closing Thoughts&#8230;</h2><p>Retrieval augmentation Generation in LLMs significantly amplifies their strengths, enhancing perplexity, accuracy, and learning capacities. The evidence is compelling: a 4K context LLM, when powered by retrieval, can go toe-to-toe with a 16K context counterpart, ensuring computational efficiency during inference.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://ai-unltd.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Enjoyed diving deep into the world of RAG and long-context LLMs? Subscribe to the <strong>AI Newsletter Today</strong> to keep up with the latest advancements in AI!</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Extending LLAMA to 32K tokens - Catching Up with ChatGPT]]></title><description><![CDATA[LLAMA 2 LONG defeats GPT-3.5-16K ChatGPT model]]></description><link>https://ai-unltd.com/p/extending-llama-to-32k-tokens-catching</link><guid isPermaLink="false">https://ai-unltd.com/p/extending-llama-to-32k-tokens-catching</guid><dc:creator><![CDATA[Arjun]]></dc:creator><pubDate>Fri, 06 Oct 2023 04:30:05 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!GjRB!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F952b7a1e-52c7-4138-970b-dc8b2ac79370_1024x768.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!GjRB!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F952b7a1e-52c7-4138-970b-dc8b2ac79370_1024x768.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!GjRB!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F952b7a1e-52c7-4138-970b-dc8b2ac79370_1024x768.jpeg 424w, https://substackcdn.com/image/fetch/$s_!GjRB!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F952b7a1e-52c7-4138-970b-dc8b2ac79370_1024x768.jpeg 848w, https://substackcdn.com/image/fetch/$s_!GjRB!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F952b7a1e-52c7-4138-970b-dc8b2ac79370_1024x768.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!GjRB!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F952b7a1e-52c7-4138-970b-dc8b2ac79370_1024x768.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!GjRB!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F952b7a1e-52c7-4138-970b-dc8b2ac79370_1024x768.jpeg" width="524" height="393" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/952b7a1e-52c7-4138-970b-dc8b2ac79370_1024x768.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:768,&quot;width&quot;:1024,&quot;resizeWidth&quot;:524,&quot;bytes&quot;:409726,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!GjRB!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F952b7a1e-52c7-4138-970b-dc8b2ac79370_1024x768.jpeg 424w, https://substackcdn.com/image/fetch/$s_!GjRB!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F952b7a1e-52c7-4138-970b-dc8b2ac79370_1024x768.jpeg 848w, https://substackcdn.com/image/fetch/$s_!GjRB!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F952b7a1e-52c7-4138-970b-dc8b2ac79370_1024x768.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!GjRB!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F952b7a1e-52c7-4138-970b-dc8b2ac79370_1024x768.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><strong>TL;DR</strong>: Meta's advancement in LLAMA 2 models manifests in handling long contexts, defeating the GPT-3.5-16K model. The newly introduced models: LLAMA 2 LONG 7B, 13B, 34B, and 70B, are centered around long-context continual pretraining, negating the need for a vast volume of long texts. The smaller 7B/13B variants were trained with 32,768-token sequences while the larger 34B/70B variants with 16,384-token sequences, demonstrated a significant performance improvement with an increased context window.</p><h3>Business Implications</h3><ul><li><p><strong>Enhanced Data Security</strong>: By self-hosting LLAMA 2 LONG models, companies can bolster data security while sustaining a model performance comparable to GPT, mitigating reliance on external providers.</p></li><li><p><strong>Cost-Effectiveness</strong>: The adoption of LLAMA 2 LONG models presents a more economical alternative to OpenAI&#8217;s GPT, with expenses relegated primarily to server hosting, thus reducing operational costs.</p></li><li><p><strong>Efficient Long-Context Handling</strong>: The proficiency of LLAMA 2 LONG in managing long-context tasks paves the way for developers to create chat solutions devoid of semantic retrievers, enhancing LLMs' capacity in question-answering tasks over extensive documents, streamlining the development process.</p></li></ul><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://ai-unltd.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://ai-unltd.com/subscribe?"><span>Subscribe now</span></a></p><div><hr></div><p>Meta's recent stride in Large Language Models (LLMs) lays down a milestone in extending the capabilities of open-source LLMs, particularly LLAMA 2, to handle long contexts efficiently, nudging closer to the prowess of models like GPT-4. There are 4 new models:</p><ul><li><p>LLAMA 2 LONG 7B</p></li><li><p>LLAMA 2 LONG 13B</p></li><li><p>LLAMA 2 LONG 34B</p></li><li><p>LLAMA 2 LONG 70B</p></li></ul><h2>Expanding the Horizons of Language Models</h2><p>Meta has taken a giant leap by introducing a series of long-context LLMs supporting effective context windows of up to 32,768 tokens. Through the lens of continual pretraining, LLAMA 2 has been extended with longer training sequences on a dataset where long texts were upsampled, starting a new frontier for open-source language models.</p><h2>Achieving Superior Performance</h2><p>Meta's research challenges the assumption that a wealth of long texts in the pre-train dataset is crucial for excelling in long-context tasks. It reveals that the method of long-context continual pretraining, not the volume of long texts, is key to achieving superior performance.</p><p>The strategy of long-context continual pretraining builds upon the existing knowledge and architecture of LLAMA 2, rather than starting from scratch with long sequences. This approach is less resource-intensive and time-efficient, showcasing notable improvement in long-context tasks. It utilizes a dataset where long texts are upsampled, providing a solid foundation for the LLAMA models to handle extended contexts efficiently.</p><p>Furthermore, the research uncovers a significant between performance improvement and context length, indicating continual performance improvement as the context length increases, up to 32,768 tokens.</p><h2>A Glimpse into the Training Arena</h2><p>The training regime embarked upon was both simple and cost-effective. A total of 400 billion tokens, formed as long training sequences, were utilized to train the existing LLAMA 2 model. The smaller 7B/13B variants were trained with 32,768-token sequences while the larger 34B/70B variants with 16,384-token sequences. The positional encoding modifications were a cornerstone to ensure the model's adeptness at attending to longer sequences.</p><h2>Instruction Tuning for Long-Context Tasks</h2><p>Instruction tuning emerged as a key ingredient in navigating the challenges of LLM alignment, especially under long-context scenarios. A simple yet effective approach leveraging a pre-built short-prompt dataset exhibited surprising efficacy on long-context benchmarks.</p><h2>Comparative Analysis and Results</h2><p>When pitted against benchmarks, the LLAMA 2 models trained with long sequences showcased significant improvements, especially on long-context tasks. The end result was a chat model, devoid of any human-annotated data, displaying a stronger overall performance than gpt-3.5-turbo-16k and other open-source models across a suite of long-context benchmarks. However, when thrown in the ring with the GPT-4 32K model, the LLAMA 2 Long 70B model fell short, indicating there's still ground to cover.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!ys_-!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8564cc05-191e-4e92-b756-0aa8f43bae51_795x166.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!ys_-!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8564cc05-191e-4e92-b756-0aa8f43bae51_795x166.png 424w, https://substackcdn.com/image/fetch/$s_!ys_-!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8564cc05-191e-4e92-b756-0aa8f43bae51_795x166.png 848w, https://substackcdn.com/image/fetch/$s_!ys_-!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8564cc05-191e-4e92-b756-0aa8f43bae51_795x166.png 1272w, https://substackcdn.com/image/fetch/$s_!ys_-!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8564cc05-191e-4e92-b756-0aa8f43bae51_795x166.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!ys_-!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8564cc05-191e-4e92-b756-0aa8f43bae51_795x166.png" width="795" height="166" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8564cc05-191e-4e92-b756-0aa8f43bae51_795x166.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:166,&quot;width&quot;:795,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:44405,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!ys_-!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8564cc05-191e-4e92-b756-0aa8f43bae51_795x166.png 424w, https://substackcdn.com/image/fetch/$s_!ys_-!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8564cc05-191e-4e92-b756-0aa8f43bae51_795x166.png 848w, https://substackcdn.com/image/fetch/$s_!ys_-!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8564cc05-191e-4e92-b756-0aa8f43bae51_795x166.png 1272w, https://substackcdn.com/image/fetch/$s_!ys_-!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8564cc05-191e-4e92-b756-0aa8f43bae51_795x166.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Source: <strong>Effective Long-Context Scaling of Foundation Models</strong></figcaption></figure></div><p><a href="https://arxiv.org/abs/2309.16039">Paper Link</a></p><h2>Bridging the Future: LLMs in Complex Use Cases and Beyond</h2><p>The narrative of LLMs is evolving rapidly with each passing day. They now stand on the verge of serving more intricate use cases, from analyzing knowledge-rich documents to powering more genuine chat interactions. This expedition of Meta reflects not just a technical advancement but a step towards a future where human-digital interactions are more intuitive and enriched.</p><p>The journey of extending LLAMA 2 to 32k tokens while keeping an eye on GPT models reflects a competitive spirit driving the field towards uncharted territories. The ingenuity in training methodologies and instruction tuning, as demonstrated, not only propels LLAMA 2 closer to the prowess of GPT but also lights the way for future endeavours in the domain.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://ai-unltd.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading the AI Newsletter Today! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Reducing Hallucinations in ChatGPT with Chain-of-Verification (CoVe)]]></title><description><![CDATA[Using Chain-of-Verification, Llama outperformed leading models like ChatGPT, InstructGPT, and PerplexityAI.]]></description><link>https://ai-unltd.com/p/reducing-hallucinations-in-chatgpt</link><guid isPermaLink="false">https://ai-unltd.com/p/reducing-hallucinations-in-chatgpt</guid><dc:creator><![CDATA[Arjun]]></dc:creator><pubDate>Mon, 25 Sep 2023 19:41:09 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!9NfE!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1afd7be3-b61b-46e7-8002-eaaedee0b8c7_2040x1152.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!9NfE!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1afd7be3-b61b-46e7-8002-eaaedee0b8c7_2040x1152.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!9NfE!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1afd7be3-b61b-46e7-8002-eaaedee0b8c7_2040x1152.jpeg 424w, https://substackcdn.com/image/fetch/$s_!9NfE!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1afd7be3-b61b-46e7-8002-eaaedee0b8c7_2040x1152.jpeg 848w, https://substackcdn.com/image/fetch/$s_!9NfE!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1afd7be3-b61b-46e7-8002-eaaedee0b8c7_2040x1152.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!9NfE!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1afd7be3-b61b-46e7-8002-eaaedee0b8c7_2040x1152.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!9NfE!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1afd7be3-b61b-46e7-8002-eaaedee0b8c7_2040x1152.jpeg" width="1456" height="822" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/1afd7be3-b61b-46e7-8002-eaaedee0b8c7_2040x1152.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:822,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1530747,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!9NfE!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1afd7be3-b61b-46e7-8002-eaaedee0b8c7_2040x1152.jpeg 424w, https://substackcdn.com/image/fetch/$s_!9NfE!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1afd7be3-b61b-46e7-8002-eaaedee0b8c7_2040x1152.jpeg 848w, https://substackcdn.com/image/fetch/$s_!9NfE!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1afd7be3-b61b-46e7-8002-eaaedee0b8c7_2040x1152.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!9NfE!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1afd7be3-b61b-46e7-8002-eaaedee0b8c7_2040x1152.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><strong>TL;DR:</strong> Hallucinations in LLMs refer to incorrect yet seemingly plausible outputs. The Chain-of-Verification (CoVe) method seeks to mitigate this by having LLMs draft, verify, and refine responses. <strong>Llama 65B, using CoVe, surpassed models like ChatGPT in long-form tasks</strong>. CoVe's efficacy was notably tested on Wikidata and other tasks, showing an improvement in precision, F1 score, and fact score. Despite its potential, CoVe still has some limitations in completely eradicating hallucinations.</p><h3>Business Implications</h3><ol><li><p>CoVe's improved LLM accuracy can lead to better AI-driven decision-making for businesses.</p></li><li><p>Enhanced trust in AI outputs can foster stronger customer relationships and brand reputation.</p></li><li><p>Llama 65B with CoVe might provide superior AI performance and security, differentiating businesses in the market.</p></li><li><p>Despite CoVe's advancements, businesses should maintain a hybrid approach, combining AI insights with human judgment.</p></li><li><p>CoVe's proficiency in short-form tasks indicates its potential for efficient chatbots and automated customer interactions.</p></li></ol><p>Here is a sample ChatGPT conversation to show the effectiveness of Chain-of-Verification (CoVe): <a href="https://chat.openai.com/share/c716c501-2d41-478a-a436-019aa378b717">Chat Link</a></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://www.ainewsletter.today/&quot;,&quot;text&quot;:&quot;Subscribe to read more such articles&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://www.ainewsletter.today/"><span>Subscribe to read more such articles</span></a></p><div><hr></div><p>Hallucinations in LLMs refer to the generation of plausible yet factually incorrect information. As LLMs are trained on an enormous text corpus, spanning billions of tokens, their performance generally improves with an increase in model parameters. Yet, even the most advanced models can falter, especially on tasks less represented in their training data. These errors often appear credible yet are factually incorrect.</p><h2>Chain-of-Verification (CoVe): A Solution to Hallucinations</h2><p>Enter the Chain-of-Verification (CoVe) method, a novel approach designed to curb the hallucination issue in LLMs. The CoVe method is a systematic process where the LLM:</p><ul><li><p>Draft an initial response</p></li><li><p>Plan verification questions to fact-check the draft</p></li><li><p>Answer the planned verification questions independently to avoid bias</p></li><li><p>Generates a final, verified response</p></li></ul><h2>Deep Dive: How CoVe Works</h2><ul><li><p><strong>Baseline Response</strong>: A simple output from LLM is obtained as the starting point, which is typically prone to hallucinations.</p></li><li><p><strong>Plan Verifications</strong>: Using the baseline response, LLM generates a series of verification questions that test the factual claims of the baseline responses.</p></li><li><p><strong>Execute Verifications</strong>: LLM answers the planned verification questions through several variations like <em>Joint, 2-Step, Factored, and Factor+Revise</em>, each with its unique approach and level of sophistication.</p></li><li><p><strong>Generate Final Verified Response</strong>: The improved response that takes verification into account is generated, incorporating any discovered inconsistencies.</p></li></ul><h2>CoVe in Action: Experimental Results</h2><p>CoVe's efficacy was tested across various tasks, including Wikidata, Wikipedia Category List, MultiSpanQA, and long-form biographies. The results are promising:</p><ul><li><p>Significant precision improvements in list-based tasks.</p></li><li><p>Enhanced performance in closed book QA, with a 23% F1 score improvement (This represents an improvement to both precision and recall).</p></li><li><p>A 28% increase in fact score for long-form generations.</p></li><li><p><strong>Notably, with CoVe, Llama 65B outperformed leading models like ChatGPT, InstructGPT, and PerplexityAI in long-form generation tasks, marking a significant achievement in the realm of open-source LLMs.</strong></p></li></ul><p><a href="https://arxiv.org/abs/2309.11495">Paper Link</a>.</p><p><a href="https://chat.openai.com/share/c716c501-2d41-478a-a436-019aa378b717">Check this ChatGPT conversation for prompt examples.</a></p><h2>Additional Insights from the Study</h2><p>The experiments also revealed that short-form verification questions were more accurately answered than long-form ones. Additionally, LLM-based verification questions surpassed heuristic-based ones, and open-ended questions proved more effective than yes/no formats.</p><h2>Limitations of the CoVe Method</h2><p>Despite its groundbreaking approach, CoVe isn't without limitations. While it significantly reduces hallucinations, it doesn't eradicate them entirely. There's still a possibility of the model generating misleading information. Moreover, hallucinations might manifest in other forms, such as during incorrect reasoning or when expressing opinions in long-form answers.</p><h2>Conclusion</h2><p>The Chain-of-Verification (CoVe) method represents a significant stride in reducing hallucinations in Large Language Models, enhancing their reliability and accuracy across various tasks. By enabling models to verify their responses, CoVe brings us closer to more dependable and error-free artificial intelligence, although some limitations and challenges still need to be addressed.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://ai-unltd.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading the AI Newsletter Today! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[PDF Triage: Elevating ChatGPT's Question Answering Capabilities]]></title><description><![CDATA[Unlocking Advanced Document Insights with Enhanced QA Techniques]]></description><link>https://ai-unltd.com/p/pdf-triage-elevating-chatgpts-question</link><guid isPermaLink="false">https://ai-unltd.com/p/pdf-triage-elevating-chatgpts-question</guid><dc:creator><![CDATA[Arjun]]></dc:creator><pubDate>Wed, 20 Sep 2023 04:30:13 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!IxM9!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb222d387-de6d-441f-acbd-01368fc8ebc6_1024x768.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!IxM9!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb222d387-de6d-441f-acbd-01368fc8ebc6_1024x768.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!IxM9!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb222d387-de6d-441f-acbd-01368fc8ebc6_1024x768.jpeg 424w, https://substackcdn.com/image/fetch/$s_!IxM9!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb222d387-de6d-441f-acbd-01368fc8ebc6_1024x768.jpeg 848w, https://substackcdn.com/image/fetch/$s_!IxM9!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb222d387-de6d-441f-acbd-01368fc8ebc6_1024x768.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!IxM9!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb222d387-de6d-441f-acbd-01368fc8ebc6_1024x768.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!IxM9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb222d387-de6d-441f-acbd-01368fc8ebc6_1024x768.jpeg" width="528" height="396" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b222d387-de6d-441f-acbd-01368fc8ebc6_1024x768.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:768,&quot;width&quot;:1024,&quot;resizeWidth&quot;:528,&quot;bytes&quot;:480239,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!IxM9!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb222d387-de6d-441f-acbd-01368fc8ebc6_1024x768.jpeg 424w, https://substackcdn.com/image/fetch/$s_!IxM9!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb222d387-de6d-441f-acbd-01368fc8ebc6_1024x768.jpeg 848w, https://substackcdn.com/image/fetch/$s_!IxM9!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb222d387-de6d-441f-acbd-01368fc8ebc6_1024x768.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!IxM9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb222d387-de6d-441f-acbd-01368fc8ebc6_1024x768.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><strong>TL;DR</strong>: PDFTriage enhances LLMs' ability to handle large documents by leveraging the Adobe Extract API. With functions such as fetch_pages and fetch_table, it addresses document structure and table reasoning questions as well. The gpt-35-turbo-0613 model, when using PDFTriage, produces answers with fewer retrieved tokens, increasing efficiency.</p><h3>Business Implications</h3><ol><li><p>By mirroring users' perceptions of documents, PDFTriage can elevate product UX, potentially leading to increased adoption and retention.</p></li><li><p>Its robust performance across varying document lengths offers businesses a versatile tool, reducing the need for multiple solutions.</p></li><li><p>Precise and efficient data extraction can provide CXOs with actionable insights, optimizing business strategies.</p></li></ol><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://ai-unltd.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://ai-unltd.com/subscribe?"><span>Subscribe now</span></a></p><div><hr></div><p>The ability to extract precise information from documents is important in the digital age. Large Language Models (LLMs) have been at the forefront of this, but they face challenges when the document's size exceeds its context length. This limitation often leads to inefficiencies in document question answering (QA).</p><h2>The Problem with Current LLM Approaches</h2><p>LLMs, despite their prowess, falter when the document doesn't fit within their context length. The prevailing solution has been to retrieve relevant contexts from the document and present them as plain text.</p><p>However, this approach overlooks the inherent structure of many documents. When users think of a PDF or a webpage, they visualize pages, tables, and sections. Representing these as mere text creates a disconnect between the user's mental model and the system's representation. This incongruity becomes glaringly evident when seemingly simple questions stump the QA system. </p><p>For instance:</p><ul><li><p>"Can you summarize the key takeaways from pages 5-7?"</p></li><li><p>"What year [in Table 3] has the maximum revenue?"</p></li></ul><p>Both questions require an understanding of the document's structure, something that plain text representation lacks.</p><h2>Introducing PDFTriage: Bridging the Gap</h2><p>Enter PDFTriage, a groundbreaking approach that enables models to retrieve context based on either the document's structure or its content. This method proves effective where traditional retrieval-augmented LLMs fall short. By giving models access to a document's structural metadata, PDFTriage can handle a variety of questions that stump plain retrieval-augmented LLMs.</p><h2>How PDFTriage Works</h2><p>The genius of PDFTriage lies in its three-step method:</p><ol><li><p><strong>Generate Document Metadata</strong>: Using the Adobe Extract API, PDFs are transformed into an HTML-like tree. This tree, rich with metadata like section titles, tables, and figures, is then parsed to extract valuable structural information.</p></li><li><p><strong>LLM-based Triage</strong>: The LLM queries the document, selecting precise content based on the question at hand.</p></li><li><p><strong>Answer Using Retrieved Content</strong>: With the relevant context retrieved, the LLM generates a comprehensive answer.</p></li></ol><p>PDFTriage employs five functions to achieve this:</p><ul><li><p><strong>fetch_pages</strong>: Retrieves text from specified pages.</p></li><li><p><strong>fetch_sections</strong>: Extracts text from a given section.</p></li><li><p><strong>fetch_table</strong>: Gathers text from a specified table caption.</p></li><li><p><strong>fetch_figure</strong>: Obtains text surrounding a particular figure caption.</p></li><li><p><strong>retrieve</strong>: Issues a natural language query over the document, fetching pertinent chunks.</p></li></ul><p><strong>These functions will be called by LLMs like GPT to synthesise various pieces of information to craft the final answer.</strong></p><h2>PDFTriage in Action: Testing and Results</h2><p>To validate PDFTriage's capabilities, a dataset comprising roughly 900 human-written questions spanning 90 documents was curated. These questions spanned categories like "document structure questions," "table reasoning questions," and even "trick questions". PDFTriage was tested on the gpt-35-turbo-0613 model.</p><p>The results were illuminating. Human evaluators consistently favoured PDFTriage over traditional retrieval methods. Specifically, PDFTriage was preferred 50.7% of the time, outperforming both Page Retrieval and Chunk Retrieval methods. </p><p>Moreover, PDFTriage showcased its efficiency by requiring fewer tokens to produce superior answers. Impressively, the length of the document had a negligible effect on PDFTriage's performance, underscoring its adaptability to both short and long documents.</p><p><a href="https://arxiv.org/abs/2309.08872">Paper Link</a></p><h2>Benefits of PDFTriage</h2><p>PDFTriage's approach aligns seamlessly with the user's perception of structured documents. By recognizing and utilizing a document's structure, it offers:</p><ul><li><p>Enhanced answer quality and accuracy.</p></li><li><p>Improved readability and informativeness.</p></li><li><p>Efficient answers with fewer retrieved tokens.</p></li><li><p>Consistent performance across varying document lengths.</p></li></ul><h2>Conclusion</h2><p>PDFTriage is set to redefine the realm of document QA. By aligning more closely with users' perceptions of structured documents, it offers a more intuitive and efficient solution to information retrieval. Its potential impact on future QA systems is immense, promising more accurate, efficient, and user-friendly outcomes.</p><p>For those keen on harnessing the power of advanced QA systems, it's time to delve deeper into PDFTriage. Its applications are vast, and its promise is undeniable. Embrace the future of structured document querying today.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://ai-unltd.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading the AI Newsletter Today! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[High-Quality Summaries with ChatGPT: Chain of Density (Prompt Included)]]></title><description><![CDATA[This is a groundbreaking method to generate increasingly dense summaries.]]></description><link>https://ai-unltd.com/p/high-quality-summaries-with-chatgpt</link><guid isPermaLink="false">https://ai-unltd.com/p/high-quality-summaries-with-chatgpt</guid><dc:creator><![CDATA[Arjun]]></dc:creator><pubDate>Sun, 17 Sep 2023 19:06:17 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!Vrhh!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd57c61f3-be93-413e-947c-521b948b54d4_1024x768.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Vrhh!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd57c61f3-be93-413e-947c-521b948b54d4_1024x768.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Vrhh!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd57c61f3-be93-413e-947c-521b948b54d4_1024x768.jpeg 424w, https://substackcdn.com/image/fetch/$s_!Vrhh!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd57c61f3-be93-413e-947c-521b948b54d4_1024x768.jpeg 848w, https://substackcdn.com/image/fetch/$s_!Vrhh!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd57c61f3-be93-413e-947c-521b948b54d4_1024x768.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!Vrhh!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd57c61f3-be93-413e-947c-521b948b54d4_1024x768.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Vrhh!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd57c61f3-be93-413e-947c-521b948b54d4_1024x768.jpeg" width="528" height="396" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d57c61f3-be93-413e-947c-521b948b54d4_1024x768.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:768,&quot;width&quot;:1024,&quot;resizeWidth&quot;:528,&quot;bytes&quot;:387342,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Vrhh!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd57c61f3-be93-413e-947c-521b948b54d4_1024x768.jpeg 424w, https://substackcdn.com/image/fetch/$s_!Vrhh!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd57c61f3-be93-413e-947c-521b948b54d4_1024x768.jpeg 848w, https://substackcdn.com/image/fetch/$s_!Vrhh!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd57c61f3-be93-413e-947c-521b948b54d4_1024x768.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!Vrhh!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd57c61f3-be93-413e-947c-521b948b54d4_1024x768.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>In today's online world, where there's so much information, turning big chunks of text into short, clear summaries is very important. Here comes ChatGPT (GPT-4). Among the many things it can do with a simple prompt, making quick summaries is one of its top skills, helping users get the main idea of long articles, reports, or data easily and quickly.</p><h2>The Challenge of Summarization</h2><p>Crafting the perfect summary is no easy feat. It involves selecting just the right amount of information&#8212;enough to be detailed and entity-centric, but not so much that it becomes dense and hard to follow. This delicate balance is what researchers from Salesforce AI, MIT, and Columbia University aimed to strike with their innovative "Chain of Density" (CoD) prompt.</p><h2>Introducing the Chain of Density (CoD) Prompting</h2><p>The CoD approach is a groundbreaking method that seeks to generate increasingly dense GPT-4 summaries. The process begins with a very simple summary. From there, GPT-4 iteratively incorporates missing important information, all without increasing the summary's length.</p><h2>How Chain of Density Works</h2><p>The CoD method is iterative. It starts with a summary that focuses on just 1-3 initial key pieces of information (entities). As the process continues, 1-3 missing entities are identified from the source text and seamlessly fused into the summary, all while maintaining the original length. This is achieved through a combination of abstraction, compression, and fusion. The aim is to convey more information within a fixed token budget, ensuring the summary remains legible and accurate. <strong>There are 5 total iterations that happen and all these iterations can happen within a single prompt.</strong></p><h2>Comparing CoD Summaries with Traditional GPT-4 Summaries</h2><p>Summaries produced using the CoD method have distinct advantages. They are more abstractive, exhibit greater fusion, and have less lead bias than those generated by a vanilla GPT-4 prompt. A human preference study, conducted on 100 CNN/DailyMail articles, revealed that humans favoured GPT-4 summaries produced using the CoD method. These summaries were almost as dense as human-written ones and made more sense than those generated by a vanilla prompt.</p><h2>Practical Application: The CoD Prompt in Action</h2><p>The Chain of Density prompt is a marvel of innovation. Here&#8217;s the exact prompt:</p><pre><code>Article: {{ ARTICLE}}

You will generate increasingly concise, entity-dense summaries of the above Article.

Repeat the following 2 steps 5 times.

Step 1. Identify 1-3 informative Entities (";" delimited) from the Article which are missing from the previously generated summary.
Step 2. Write a new, denser summary of identical length which covers every entity and detail from the previous summary plus the Missing Entities.

A Missing Entity is:
- Relevant: to the main story.
- Specific: descriptive yet concise (5 words or fewer).
- Novel: not in the previous summary.
- Faithful: present in the Article.
- Anywhere: located anywhere in the Article.

Guidelines:
- The first summary should be long (4-5 sentences, ~80 words) yet highly non-specific, containing little information beyond the entities marked as missing. Use overly verbose language and fillers (e.g., "this article discusses") to reach ~80 words.
- Make every word count: re-write the previous summary to improve flow and make space for additional entities.
- Make space with fusion, compression, and removal of uninformative phrases like "the article discusses".
- The summaries should become highly dense and concise yet self-contained, e.g., easily understood without the Article.
- Missing entities can appear anywhere in the new summary.
- Never drop entities from the previous summary. If space cannot be made, add fewer new entities. Remember, use the exact same number of words for each summary. 

Answer in JSON. The JSON should be a list ( length 5) of dictionaries whose keys are "Missing_Entities" and "Denser_Summary"</code></pre><p><em>Tip: Include the article in XML tags such as </em><code>&lt;Article&gt;&lt;/Article&gt;</code><em> for good results</em></p><p>When applied to an article titled "<a href="https://edition.cnn.com/cnn-underscored/electronics/iphone-15-hands-on">iPhone 15 hands-on: pre-orders, release date, price</a>", the following output was produced:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!v7l7!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F870f56ab-0833-4fa0-a3f9-80a75c425b8d_876x608.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!v7l7!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F870f56ab-0833-4fa0-a3f9-80a75c425b8d_876x608.png 424w, https://substackcdn.com/image/fetch/$s_!v7l7!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F870f56ab-0833-4fa0-a3f9-80a75c425b8d_876x608.png 848w, https://substackcdn.com/image/fetch/$s_!v7l7!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F870f56ab-0833-4fa0-a3f9-80a75c425b8d_876x608.png 1272w, https://substackcdn.com/image/fetch/$s_!v7l7!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F870f56ab-0833-4fa0-a3f9-80a75c425b8d_876x608.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!v7l7!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F870f56ab-0833-4fa0-a3f9-80a75c425b8d_876x608.png" width="876" height="608" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/870f56ab-0833-4fa0-a3f9-80a75c425b8d_876x608.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:608,&quot;width&quot;:876,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:124041,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!v7l7!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F870f56ab-0833-4fa0-a3f9-80a75c425b8d_876x608.png 424w, https://substackcdn.com/image/fetch/$s_!v7l7!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F870f56ab-0833-4fa0-a3f9-80a75c425b8d_876x608.png 848w, https://substackcdn.com/image/fetch/$s_!v7l7!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F870f56ab-0833-4fa0-a3f9-80a75c425b8d_876x608.png 1272w, https://substackcdn.com/image/fetch/$s_!v7l7!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F870f56ab-0833-4fa0-a3f9-80a75c425b8d_876x608.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><strong>The 3rd output seems to be the best.</strong></p><p><a href="https://arxiv.org/abs/2309.04269">Paper Link</a></p><h2>Evaluating the Quality of CoD Summaries</h2><p>The quality of CoD summaries was put to the test through both human and GPT-4 evaluations. Human annotators were presented with randomly shuffled CoD summaries from 100 CNN/DailyMail articles. Their feedback was enlightening: for 3 out of 4 annotators, <strong>the first iteration of CoD received the most first-place votes across the 100 examples. However, in aggregate, 61% of top-ranked summaries had undergone at least three iterations</strong>. This suggests that a significant portion of annotators favoured summaries that had undergone further densification. The inferred entity density, calculated as the ratio of entities mentioned per token, after three iterations was approximately 0.15, closely mirroring human-written summaries (0.151) and surpassing those produced by a vanilla GPT-4 prompt.</p><p>GPT-4 itself showcased its versatility by being able to evaluate the quality of summaries. When prompted, it provided ratings, revealing a <strong>preference for the middle iterations with scores of 4.78, 4.77, and 4.76, while the first and last iterations were less favoured</strong>. This dual approach to evaluation, combining human insights with machine precision and specific ratings, offers a comprehensive understanding of summary quality.</p><h2>Limitations and Future Directions</h2><p>While the CoD method has shown promise, it's worth noting that its analysis has been limited to news summarization. Additionally, CoD has been exclusively tested on GPT-4, a closed-source model, and hasn't been evaluated on other large language models (LLMs). There's potential for its application across various domains, and the quest continues to determine the optimal density for summaries. The broader applicability of CoD on different LLMs remains an area ripe for exploration.</p><h2>Conclusion</h2><p>The Chain of Density prompting method offers a fresh perspective on automated text summarization. It's an invitation for readers and researchers alike to experiment with CoD and GPT-4, pushing the boundaries of what's possible in the realm of concise, high-quality summaries.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://ai-unltd.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Enjoyed diving deep into the world of GPT-4 and the Chain of Density Prompting? Subscribe to <strong>AI Newsletter Today</strong> to keep up with the latest advancements in AI!</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Automatic Audiobook Creation Using Neural Text-To-Speech]]></title><description><![CDATA[Over 5000 audiobooks were generated using Neural Text-To-Speech and Project Gutenberg e-books.]]></description><link>https://ai-unltd.com/p/automatic-audiobook-creation-using</link><guid isPermaLink="false">https://ai-unltd.com/p/automatic-audiobook-creation-using</guid><dc:creator><![CDATA[Arjun]]></dc:creator><pubDate>Thu, 14 Sep 2023 04:31:00 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!cfU2!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd98de9d0-9e77-4f1b-b9b1-86214b102a91_1024x768.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!cfU2!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd98de9d0-9e77-4f1b-b9b1-86214b102a91_1024x768.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!cfU2!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd98de9d0-9e77-4f1b-b9b1-86214b102a91_1024x768.jpeg 424w, https://substackcdn.com/image/fetch/$s_!cfU2!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd98de9d0-9e77-4f1b-b9b1-86214b102a91_1024x768.jpeg 848w, https://substackcdn.com/image/fetch/$s_!cfU2!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd98de9d0-9e77-4f1b-b9b1-86214b102a91_1024x768.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!cfU2!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd98de9d0-9e77-4f1b-b9b1-86214b102a91_1024x768.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!cfU2!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd98de9d0-9e77-4f1b-b9b1-86214b102a91_1024x768.jpeg" width="560" height="420" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d98de9d0-9e77-4f1b-b9b1-86214b102a91_1024x768.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:768,&quot;width&quot;:1024,&quot;resizeWidth&quot;:560,&quot;bytes&quot;:492223,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!cfU2!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd98de9d0-9e77-4f1b-b9b1-86214b102a91_1024x768.jpeg 424w, https://substackcdn.com/image/fetch/$s_!cfU2!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd98de9d0-9e77-4f1b-b9b1-86214b102a91_1024x768.jpeg 848w, https://substackcdn.com/image/fetch/$s_!cfU2!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd98de9d0-9e77-4f1b-b9b1-86214b102a91_1024x768.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!cfU2!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd98de9d0-9e77-4f1b-b9b1-86214b102a91_1024x768.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><strong>TL;DR:</strong> Researchers from Microsoft, MIT, Project Gutenberg and Google have collaborated to revolutionize audiobook creation. Using neural text-to-speech technology and the SynapseML framework, they've automated the process, turning e-books into high-quality, customizable audiobooks. This system overcomes previous robotic voice challenges, offers voice customization, and introduces emotive reading, making literature more accessible and engaging.</p><p>Here&#8217;s an AI-generated audiobook:</p><iframe class="spotify-wrap podcast" data-attrs="{&quot;image&quot;:&quot;https://i.scdn.co/image/ab6765630000ba8ab5b43154e3b1dc9ff7dfa08e&quot;,&quot;title&quot;:&quot;Trent's Trust, and Other Stories by Bret Harte&quot;,&quot;subtitle&quot;:&quot;Project Gutenberg&quot;,&quot;description&quot;:&quot;Episode&quot;,&quot;url&quot;:&quot;https://open.spotify.com/episode/2mFIf7SFpoY2JP8LRbjN0K&quot;,&quot;belowTheFold&quot;:false,&quot;noScroll&quot;:false}" src="https://open.spotify.com/embed/episode/2mFIf7SFpoY2JP8LRbjN0K" frameborder="0" gesture="media" allowfullscreen="true" allow="encrypted-media" data-component-name="Spotify2ToDOM"></iframe><div><hr></div><p>In the digital age, audiobooks have emerged as a pivotal medium, enhancing the accessibility and engagement of literature for audiences worldwide. However, the traditional methods of creating these audiobooks are fraught with challenges, from the extensive time and effort required to the inconsistencies in quality. Enter the groundbreaking collaboration between researchers from tech and academic giants: Microsoft, MIT, Project Gutenberg, and Google. Together, they're reshaping the audiobook landscape, introducing automation and innovation where it's needed most.</p><p><a href="https://marhamilresearch4.blob.core.windows.net/gutenberg-public/Website/index.html#Code">Website Link</a></p><p><a href="https://arxiv.org/abs/2309.03926">Paper Link</a></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://ai-unltd.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://ai-unltd.com/subscribe?"><span>Subscribe now</span></a></p><h2>The Need for Automated Audiobook Creation</h2><p>Historically, producing an audiobook has been a labour-intensive process. Whether it's the meticulous narration by professionals or the passionate efforts of volunteers, the journey from text to audio is long and winding. Platforms like LibriVox, driven by human volunteers, have made commendable strides in making audiobooks accessible. However, the variability in recording quality and environments can lead to inconsistent outputs. On the other end of the spectrum, platforms like Audible offer high-quality audiobooks but at a price, both monetarily and in terms of open access.</p><h2>Unveiling the Automated System</h2><p>The collaborative project introduces a system that harnesses the power of neural text-to-speech technology, promising to revolutionize the way we perceive audiobooks. Imagine generating thousands of human-quality audiobooks from the vast collection of Project Gutenberg, all at the click of a button. Beyond this, the system offers unparalleled customization. Listeners can adjust speaking speeds, choose different styles, and even modify emotional intonations. And for those who've always dreamt of hearing a book in their voice, this system can make that dream a reality with just a snippet of sample audio.</p><h2>Overcoming Traditional Challenges</h2><p>Past attempts at automated audiobook creation often stumbled at two major hurdles: the unmistakably robotic tone of text-to-speech systems and the challenge of discerning which text segments should be vocalized (you don&#8217;t want a huge index table readout in an audiobook!). This new system not only produces high-quality, human-like audio but also intelligently navigates the content of diverse e-books, ensuring a seamless listening experience.</p><h2>The Technical Backbone: SynapseML</h2><p>At the heart of this revolution is SynapseML, a robust scalable machine learning framework. It orchestrates the entire audiobook creation process, ensuring efficiency and quality at every step. From parsing thousands of e-books from Project Gutenberg to generating emotive, lifelike speech, SynapseML is the unsung hero behind the scenes.</p><h2>From E-books to Audiobooks: The Process</h2><p>The journey begins with e-books, specifically those in the HTML format from Project Gutenberg. Given the non-standardized nature of these files, which can contain everything from footnotes to transcriber notes, the system employs clustering on the HTML Document Object Model (DOM) tree. This approach allows for the efficient parsing of vast collections, ensuring only the relevant text makes its way to the listener's ears.</p><h2>Achieving High-Quality Speech</h2><p>The system's prowess doesn't stop at parsing. It recognizes the nuances required in different genres. While a non-fiction work might demand a clear, neutral voice, fiction, with its dialogues and drama, comes alive with emotive reading. Using the zero-shot text-to-speech method, it can even clone a user's voice from minimal recordings, adding a personal touch to the listening experience.</p><h2>Breathing Life into Text: Emotive Reading</h2><p>What truly sets this system apart is its ability to infuse emotion into the narrative. By segmenting text into narration and dialogue, identifying speakers, and predicting emotions, it crafts a dynamic listening experience. Passages with multiple characters and emotional dialogues are no longer monotonous; they're vibrant and engaging, much like a theatrical performance.</p><h2>Conclusion</h2><p>The collaboration between researchers of Microsoft, MIT, Project Gutenberg, and Google marks a significant leap in the world of audiobooks. By automating the creation process, they're not only making literature more accessible but also ensuring consistent quality. As we stand on the cusp of this revolution, one thing is clear: the future of audiobooks is bright, inclusive, and incredibly exciting.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://ai-unltd.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading the AI Newsletter Today! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[ChatGPT Can Build a Complete Software in Less Than 7 Minutes: ChatDev]]></title><description><![CDATA[Dive into the world of ChatDev, and experience the future of software development.]]></description><link>https://ai-unltd.com/p/chatgpt-can-build-a-software-development</link><guid isPermaLink="false">https://ai-unltd.com/p/chatgpt-can-build-a-software-development</guid><dc:creator><![CDATA[Arjun]]></dc:creator><pubDate>Thu, 07 Sep 2023 04:30:08 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!bg8e!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8416dbc8-6313-4923-a70e-98f3d49e3eba_1024x768.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!bg8e!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8416dbc8-6313-4923-a70e-98f3d49e3eba_1024x768.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!bg8e!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8416dbc8-6313-4923-a70e-98f3d49e3eba_1024x768.jpeg 424w, https://substackcdn.com/image/fetch/$s_!bg8e!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8416dbc8-6313-4923-a70e-98f3d49e3eba_1024x768.jpeg 848w, https://substackcdn.com/image/fetch/$s_!bg8e!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8416dbc8-6313-4923-a70e-98f3d49e3eba_1024x768.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!bg8e!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8416dbc8-6313-4923-a70e-98f3d49e3eba_1024x768.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!bg8e!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8416dbc8-6313-4923-a70e-98f3d49e3eba_1024x768.jpeg" width="570" height="427.5" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8416dbc8-6313-4923-a70e-98f3d49e3eba_1024x768.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:768,&quot;width&quot;:1024,&quot;resizeWidth&quot;:570,&quot;bytes&quot;:468459,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!bg8e!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8416dbc8-6313-4923-a70e-98f3d49e3eba_1024x768.jpeg 424w, https://substackcdn.com/image/fetch/$s_!bg8e!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8416dbc8-6313-4923-a70e-98f3d49e3eba_1024x768.jpeg 848w, https://substackcdn.com/image/fetch/$s_!bg8e!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8416dbc8-6313-4923-a70e-98f3d49e3eba_1024x768.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!bg8e!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8416dbc8-6313-4923-a70e-98f3d49e3eba_1024x768.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The process of building software is often seen as a lengthy and complex endeavour, filled with layers of decisions, consultations, and nuanced intuition. Enter ChatDev, an innovative paradigm powered by ChatGPT, which promises to flip this narrative, presenting a virtual software development company that operates at lightning speed and minimal cost.</p><h2>The ChatDev Framework</h2><p>ChatDev mirrors an entire software development company, it offers a comprehensive approach to building software applications. The secret sauce? A structured waterfall model that segregates the software creation process into distinct phases. Adding another layer of effectiveness, the chat chain deconstructs these phases into granular subtasks, ensuring precision and coherence at every step.</p><p>A standout feature of ChatDev is its innovative use of personas, all simulated by Large Language Models (LLMs). These personas, ranging from CEOs to programmers, art designers to testers, embody the roles of a real-world software development team. <strong>Each persona has its unique expertise and perspective, bringing depth and diversity to the decision-making process</strong>. They interact, collaborate, and debate just as human counterparts would in a traditional software company. This multi-agent approach, entirely powered by ChatGPT, creates a dynamic and responsive virtual development environment.</p><h2>The Four Phases of ChatDev</h2><h4>1. Designing</h4><p>The journey begins with the 'Designing' phase. Here, innovative ideas are crafted. The CEO, CPO, and CTO personas (all played by ChatGPT agents) bring their expertise to the table, turning abstract concepts into well-defined technical designs. They discuss and decide on the software's modality and the ideal programming language to bring it to life.</p><h4>2. Coding</h4><p>With a design in hand, the 'Coding' phase takes centre stage. The CTO collaborates with the programmer and the art designer, translating designs into functional code and captivating user interfaces. Every line of code, every design element, is meticulously crafted in this collaborative endeavour.</p><h4>3. Testing</h4><p>But what's coding without rigorous testing? In the 'Testing' phase, the software undergoes thorough scrutiny. The programmer, reviewer, and tester personas ensure that the code doesn't just work; it excels. From peer reviews to system testing, every potential glitch is identified and rectified.</p><h4>4. Documenting</h4><p>The final stage in this journey is 'Documenting'. This isn't just about creating a user manual; it's about crafting a comprehensive guide that details environment specifications and everything a user needs to make the most of the software. The CEO, CPO, CTO, and programmer unite to ensure that users have clarity from the get-go.</p><h2>Addressing the Hallucination Challenge</h2><p>Hallucinations in ChatGPT agents could lead to incomplete functions, missing dependencies, or even undiscovered bugs. However, ChatDev's granular task approach and cross-examination mechanism serve as a safeguard against these hallucinations, ensuring clarity and precision.</p><h2>The Role of Communication in ChatDev</h2><p>Communication is the heart of ChatDev. The framework thrives on context-aware, multi-turn discussions, ensuring that every decision, from software design to bug fixes, is a result of collaborative dialogue. This context-rich communication is further enhanced by the thought instruction mechanism.</p><p>The thought instruction mechanism introduces a "role flip" during the code completion, reviewing, and testing stages. An instructor agent provides specific guidelines for code modifications, directing the assistant programmer agent. This ensures precision and alignment with desired outputs, minimizing errors and enhancing the overall quality of the software being developed.</p><h2>Experiments, Results, and Findings</h2><p>Rigorous experiments using the GPT-3.5 Turbo version of ChatGPT validate ChatDev's prowess. Not only does it highlight the framework's efficacy, but it also showcases tangible metrics of its efficiency. ChatDev impressively produces software solutions for 70 unique user requirements, generating an average of 17.04 files per software. <strong>The speed is unmatched, wrapping up the entire process in just 409.84 seconds (that's under 7 minutes &#128561;) and at a staggering low cost of approximately $0.2967 (&#129327;)</strong>. These metrics elucidate the unparalleled efficiency and cost-effectiveness of the ChatDev approach.</p><p><a href="https://arxiv.org/abs/2307.07924v3?s=09">Paper Link</a></p><p><a href="https://github.com/OpenBMB/ChatDev">Code Link</a></p><h2>Conclusion</h2><p>ChatDev is more than just a software development tool; it's a testament to the future of software engineering. By harnessing the power of ChatGPT, ChatDev offers a glimpse into a world where building software is not just fast and affordable but also efficient and precise. With its structured approach, collaborative communication, and unmatched speed, ChatDev promises to reshape the software development landscape, one line of code at a time.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://ai-unltd.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading the AI Newsletter Today! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Fact Checking ChatGPT: Using FacTool to Detect Factual Errors]]></title><description><![CDATA[FacTool - A Framework Designed to Detect Factual Errors in LLM Outputs.]]></description><link>https://ai-unltd.com/p/fact-checking-chatgpt-using-factool</link><guid isPermaLink="false">https://ai-unltd.com/p/fact-checking-chatgpt-using-factool</guid><dc:creator><![CDATA[Arjun]]></dc:creator><pubDate>Wed, 06 Sep 2023 04:31:02 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!g2ML!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ee24f39-ff6a-4064-a693-4521da35d71d_1024x768.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!g2ML!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ee24f39-ff6a-4064-a693-4521da35d71d_1024x768.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!g2ML!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ee24f39-ff6a-4064-a693-4521da35d71d_1024x768.jpeg 424w, https://substackcdn.com/image/fetch/$s_!g2ML!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ee24f39-ff6a-4064-a693-4521da35d71d_1024x768.jpeg 848w, https://substackcdn.com/image/fetch/$s_!g2ML!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ee24f39-ff6a-4064-a693-4521da35d71d_1024x768.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!g2ML!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ee24f39-ff6a-4064-a693-4521da35d71d_1024x768.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!g2ML!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ee24f39-ff6a-4064-a693-4521da35d71d_1024x768.jpeg" width="542" height="406.5" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/2ee24f39-ff6a-4064-a693-4521da35d71d_1024x768.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:768,&quot;width&quot;:1024,&quot;resizeWidth&quot;:542,&quot;bytes&quot;:477668,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!g2ML!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ee24f39-ff6a-4064-a693-4521da35d71d_1024x768.jpeg 424w, https://substackcdn.com/image/fetch/$s_!g2ML!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ee24f39-ff6a-4064-a693-4521da35d71d_1024x768.jpeg 848w, https://substackcdn.com/image/fetch/$s_!g2ML!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ee24f39-ff6a-4064-a693-4521da35d71d_1024x768.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!g2ML!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F2ee24f39-ff6a-4064-a693-4521da35d71d_1024x768.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The rise of Large Language Models (LLMs) like ChatGPT has revolutionized the AI domain. These models offer high-quality text outputs, but they're not without challenges. Key among these is the issue of factual errors in generated texts. This article delves into the world of FacTool, a beacon in the murky waters of LLM-generated content.</p><h2>Challenges and Limitations of LLM-Generated Texts</h2><p>LLMs have ushered in a new age of content creation, handling a vast array of tasks from question answering to code generation. However, with their prowess come certain limitations:</p><ul><li><p>A heightened risk of factual inconsistencies across diverse tasks.</p></li><li><p>Outputs that, while detailed, often blur individual factual boundaries.</p></li><li><p>A notable absence of concrete sources/evidence accompanying the generated content.</p></li><li><p>An inherent tendency to produce text that, while sounding credible, might be riddled with inaccuracies.</p></li></ul><p>These limitations are especially critical in high-stakes domains like healthcare, finance, and law, emphasizing the dire need for rigorous fact-checking.</p><h2>The FacTool Solution</h2><p>Addressing the above challenges is FacTool &#8212; a domain-agnostic framework designed to detect factual errors in texts produced by LLMs. Think of FacTool as a guardian, ensuring the veracity of every piece of information generated by models like ChatGPT.</p><h2>Tested Domains for FacTool</h2><p>FacTool's versatility shines through its testing across multiple tasks:</p><ul><li><p><strong>Knowledge-based QA</strong>: Validating the accuracy of answers.</p></li><li><p><strong>Code Generation</strong>: Validating the efficacy of generated codes.</p></li><li><p><strong>Mathematical Reasoning</strong>: Ensuring accurate problem-solving.</p></li><li><p><strong>Scientific Literature Review</strong>: Validating the credibility of AI-generated citations and reviews.</p></li></ul><h2>How FacTool Works</h2><p>Harnessing the power of tool augmentation, FacTool operates through:</p><ol><li><p><strong>Claim extraction</strong>: LLMs, like ChatGPT, extract claims from the generated response. Claims are the sentences that need to be fact-checked. For instance, in Knowledge-based QA, each claim is a part of the answer generated, while in code generation, every snippet becomes a claim to verify.</p></li><li><p><strong>Query generation</strong>: These claims are transformed into queries that can be sent to external tools to retrieve factual responses.</p></li><li><p><strong>Tool querying &amp; Evidence collection</strong>: Relevant evidence is collated by sending the generated query to external tools such as Google Search, Python interpreters or Google Scholar APIs.</p></li><li><p><strong>Agreement verification</strong>: The final step involves presenting the claim and evidence back to the LLM, which then verifies the claim's authenticity.</p></li></ol><h2>FacTool's Stellar Performance</h2><p>Benchmarked against established methods like 3-shot chain-of-thought, FacTool, especially when powered by GPT-4, consistently outperformed competitors. Key findings include:</p><ul><li><p>Superior performance in scientific literature review tasks.</p></li><li><p>Enhanced sensitivity in error detection compared to self-check chain-of-thought methods.</p></li><li><p>In Knowledge-based QA, FacTool debunked false claims, such as "Argentina has not won the World Cup since 1986".</p></li><li><p>In code generation, it surpassed other baselines, showcasing its capability in technical domains.</p></li><li><p>Both GPT-4 and ChatGPT versions of FacTool excelled in mathematical reasoning, identifying errors in calculations.</p></li></ul><p><a href="https://arxiv.org/abs/2307.13528v2">Paper link</a></p><p><a href="https://github.com/GAIR-NLP/factool">Code link</a></p><h2>Conclusion</h2><p>With the world leaning heavily on LLMs for varied tasks, the importance of tools like FacTool cannot be overstated. Ensuring the factuality of generated content paves the way for a future where we can trust AI-generated outputs without second-guessing.</p><p>For those venturing into the world of LLMs, integrating FacTool can be a game-changer. Especially in critical domains, it's more than a tool&#8212;it's a necessity. As generative AI continues to evolve, ensuring the accuracy of generated content will remain paramount, and FacTool is leading the charge in this endeavour.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://ai-unltd.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading the AI Newsletter Today! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p>]]></content:encoded></item><item><title><![CDATA[AI Roundup: ChatGPT Enterprise, Google's SynthID, Recommendation, Custom Chatbot with GPT-3.5 and more...]]></title><description><![CDATA[AI Roundup of Week 35, 2023]]></description><link>https://ai-unltd.com/p/ai-roundup-chatgpt-enterprise-googles</link><guid isPermaLink="false">https://ai-unltd.com/p/ai-roundup-chatgpt-enterprise-googles</guid><dc:creator><![CDATA[Arjun]]></dc:creator><pubDate>Mon, 04 Sep 2023 09:08:55 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe60b22df-a788-4319-8dc6-dc7b9e537ee9_256x256.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>In this weekly roundup, let&#8217;s take a look at:</p><ol><li><p><strong><a href="https://www.ainewsletter.today/i/136711520/simplest-method-to-build-a-custom-chatbot-with-gpt">Simplest Method to Build a Custom Chatbot with GPT-3.5</a></strong></p></li><li><p><strong><a href="https://www.ainewsletter.today/i/136711520/recmind-personalised-recommendation-using-chatgpt">RecMind: Personalised Recommendation using ChatGPT</a></strong></p></li><li><p><strong><a href="https://www.ainewsletter.today/i/136711520/googles-solution-to-detect-ai-generated-images">Google's Solution to Detect AI-Generated Images</a></strong></p></li><li><p><strong><a href="https://www.ainewsletter.today/i/136711520/making-chatgpt-think-like-a-human">Making ChatGPT Think Like a Human</a></strong></p></li><li><p><strong>U<a href="https://www.ainewsletter.today/i/136711520/using-ai-to-build-ai-promptmodel">sing AI to Build AI: Prompt2Model</a></strong></p></li><li><p><strong><a href="https://www.ainewsletter.today/i/136711520/using-chatgpt-in-classrooms">Using ChatGPT in Classrooms</a></strong></p></li><li><p><strong><a href="https://www.ainewsletter.today/i/136711520/chatgpt-enterprise">ChatGPT Enterprise</a></strong></p></li></ol><div><hr></div><h2>Simplest Method to Build a Custom Chatbot with GPT-3.5</h2><p>Chatbots powered by Large Language Models (LLM) are reshaping our interaction with technology. Many businesses now use them for customer support. Interestingly, the domain chat.com was sold for over $10M, showing the value of chat interfaces. These chatbots, thanks to advancements in AI, can handle intricate questions, mimicking human conversation. In this article, we explore a new tool that makes building such chatbots easier, leveraging the OpenAI GPT-3.5 Turbo model.</p><p>Read more <a href="https://www.ainewsletter.today/p/simplest-method-to-build-a-custom">here</a>.</p><div><hr></div><h2>RecMind: Personalised Recommendation using ChatGPT</h2><p>Large Language Models (LLMs) have made a mark by showcasing abilities to execute complex tasks, from solving math problems to sparking creative writing. Yet, they falter when faced with personalized queries, especially recommendation requests. Enter RecMind, an innovative research born from the collaboration between Amazon Alexa AI and Arizona State University. This LLM-powered autonomous recommender agent is tailored to address this glaring gap.</p><p>Read more <a href="https://www.ainewsletter.today/p/recmind-personalised-recommendation">here</a>.</p><div><hr></div><h2>Google's Solution to Detect AI-Generated Images</h2><p>In today's digital era, the boundary between reality and artificiality is rapidly blurring. AI-generated images, fueled by models such as stable diffusion, mid-journey, Google's Bard, and OpenAI's DALL-E 2, are gaining traction. But how do we discern their authenticity, especially when they appear convincingly realistic?</p><p>Read more <a href="https://www.ainewsletter.today/p/googles-solution-to-detect-ai-generated">here</a>.</p><div><hr></div><h2>Making ChatGPT Think Like a Human</h2><p>The landscape of the Large Language Model (LLM) prompting frameworks has seen a new contender &#8211; the Graph of Thoughts (GoT) framework. Going beyond the capabilities of Chain-of-thought (CoT) and Tree-of-thoughts (ToT), GoT presents a fresh perspective on how we can make LLMs think like humans for better results.</p><p>Read more <a href="https://www.ainewsletter.today/p/making-chatgpt-think-like-a-human">here</a>.</p><div><hr></div><h1>Using AI to Build AI: Prompt2Model</h1><p>Prompt2Model heralds a paradigm shift, underscoring the potential of using AI to construct AI models. As we stand on the cusp of this AI revolution, it's exhilarating to envision the future prospects, improvements, and the transformative impact Prompt2Model might usher in for the NLP and broader AI domain.</p><p>Read more <a href="https://www.ainewsletter.today/p/using-ai-to-build-ai-prompt2model">here</a>.</p><div><hr></div><h2>Using ChatGPT in Classrooms</h2><p>OpenAI Shared a guide for teachers on using ChatGPT in classrooms. It contains a few short stories on how educators are already using ChatGPT to accelerate student learning. It also contains prompt examples that educators can use.</p><p>Read more <a href="https://openai.com/blog/teaching-with-ai">here</a>.</p><div><hr></div><h2>ChatGPT Enterprise</h2><p>OpenAI released ChatGPT enterprise last week. This version does not use the customer prompts and company data to train GPT models. It is also SOC2 compliant. Aiming at large organizations it also has features such as SSO, custom domain, analytics dashboard, and an admin console with bulk member management.</p><p>This is an attempt by OpenAI to make large organizations comfortable in using closed-source LLMs.</p><p>Read more <a href="https://openai.com/blog/introducing-chatgpt-enterprise">here</a>.</p><div><hr></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://ai-unltd.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading the AI Newsletter Today! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Using AI to Build AI: Prompt2Model]]></title><description><![CDATA[Building deployable NLP models using GPT-3.5 Turbo and Prompt2Model Framework.]]></description><link>https://ai-unltd.com/p/using-ai-to-build-ai-prompt2model</link><guid isPermaLink="false">https://ai-unltd.com/p/using-ai-to-build-ai-prompt2model</guid><dc:creator><![CDATA[Arjun]]></dc:creator><pubDate>Sat, 02 Sep 2023 04:30:09 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!hhDS!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6815cac-2000-441d-b589-de9d86a8209e_1024x768.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!hhDS!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6815cac-2000-441d-b589-de9d86a8209e_1024x768.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!hhDS!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6815cac-2000-441d-b589-de9d86a8209e_1024x768.jpeg 424w, https://substackcdn.com/image/fetch/$s_!hhDS!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6815cac-2000-441d-b589-de9d86a8209e_1024x768.jpeg 848w, https://substackcdn.com/image/fetch/$s_!hhDS!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6815cac-2000-441d-b589-de9d86a8209e_1024x768.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!hhDS!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6815cac-2000-441d-b589-de9d86a8209e_1024x768.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!hhDS!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6815cac-2000-441d-b589-de9d86a8209e_1024x768.jpeg" width="728" height="546" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e6815cac-2000-441d-b589-de9d86a8209e_1024x768.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:false,&quot;imageSize&quot;:&quot;normal&quot;,&quot;height&quot;:768,&quot;width&quot;:1024,&quot;resizeWidth&quot;:728,&quot;bytes&quot;:424086,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!hhDS!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6815cac-2000-441d-b589-de9d86a8209e_1024x768.jpeg 424w, https://substackcdn.com/image/fetch/$s_!hhDS!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6815cac-2000-441d-b589-de9d86a8209e_1024x768.jpeg 848w, https://substackcdn.com/image/fetch/$s_!hhDS!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6815cac-2000-441d-b589-de9d86a8209e_1024x768.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!hhDS!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe6815cac-2000-441d-b589-de9d86a8209e_1024x768.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Building an AI model from the ground up can be tough. It includes deciding on the job the model will do, finding the right data, picking a model, training it, checking how good it is, and then using it. However, with the rise of Large Language Models (LLMs) like GPT-3, we can build rapid prototypes without diving deep into the coding intricacies. But is relying solely on LLMs the silver bullet we've been waiting for? Unfortunately, while LLMs promise convenience, they come with their own set of challenges, from escalating costs and slower predictions to potential privacy pitfalls. To counter this, Prompt2Model proposes that we build smaller models for specific tasks that need to be accomplished.</p><h2>The Promise of Prompt2Model</h2><p>Enter Prompt2Model &#8211; an innovative method designed to convert task descriptions into specialized, deployable models. This AI-driven tool is not just a mechanism for swift NLP system creation, but it also stands as a comprehensive research platform. Researchers can explore avenues in model distillation, synthetic evaluation, dataset retrieval, and more, all under the Prompt2Model umbrella.</p><h2>How Prompt2Model Works: An Overview</h2><p>At its core, Prompt2Model operates through a multi-pronged strategy. It harnesses existing datasets and pre-trained models, generates datasets with the assistance of LLMs like GPT-3.5 Turbo, and fine-tunes through supervised methodologies. What&#8217;s truly remarkable is its prowess. Tests have shown that Prompt2Model, with just a few-shot prompt as input, can outdo GPT-3.5 Turbo's results by an impressive average of 20%. This performance spike is achieved even when the resultant Prompt2Model is up to 700 times smaller in scale.</p><p>Note: The model that Prompt2Model builds is fine-tuned for a single task that it needs to accomplish while GPT-3.5 Turbo can accomplish multiple tasks.</p><h2>Delving Deeper: The Prompt2Model Framework</h2><p>Understanding Prompt2Model requires a dive into its intricate framework:</p><ul><li><p><strong>Prompt Parser</strong>: Here, user prompts, which could be simple task instructions or demonstrations, are decoded into tangible, actionable steps. This process ensures that the ML pipeline is optimized for user input.</p></li><li><p><strong>Dataset Retriever</strong>: With the support of DataFinder, this module scours the vast expanse of the digital realm to retrieve datasets that resonate with the task at hand.</p></li><li><p><strong>Dataset Generator</strong>: Recognizing that not all tasks are backed by existing data, this tool creates synthetic data using GPT-3.5 Turbo.</p></li><li><p><strong>Model Retriever</strong>: This module pulls an apt pre-trained model from the treasure trove of Hugging Face, ensuring the model aligns with the user's intent.</p></li><li><p><strong>Model Trainer</strong>: The chosen model undergoes fine-tuning, leveraging both the retrieved and the freshly minted datasets.</p></li><li><p><strong>Model Evaluator</strong>: Post-training, it's time for a performance check. This segment evaluates the model's accuracy and reliability.</p></li><li><p><strong>Demo Creator</strong>: An optional yet valuable feature, this tool crafts a graphical interface, allowing users to interact seamlessly with the trained model.</p></li></ul><p><a href="https://arxiv.org/abs/2308.12261?s=09">Paper Link</a></p><p><a href="https://github.com/neulab/prompt2model">Code Link</a></p><h2>The Real-World Benefits of Prompt2Model</h2><p>Prompt2Model isn't just a theoretical marvel, it promises tangible benefits. By navigating beyond the constraints of zero-shot and few-shot prompting, it delivers robust performance. In certain tasks, it even managed to overshadow GPT-3.5-turbo by an impressive average of 20%, showcasing its potential for real-world applications, especially given its efficiency and compact size.</p><h2>Limitations and Challenges</h2><p>However, Prompt2Model has a few limitations and challenges as well. Its reliance on GPT-3.5 Turbo, a paid and closed-source entity, raises eyebrows, especially when legal concerns come into play. OpenAI's policies might restrict Prompt2Model's commercial exploits. Additionally, there's a language barrier. Currently, tasks beyond the English realm might encounter hiccups, given GPT-3.5 Turbo's predominant English training.</p><h2>Conclusion</h2><p>Prompt2Model heralds a paradigm shift, underscoring the potential of using AI to construct AI models. As we stand on the cusp of this AI revolution, it's exhilarating to envision the future prospects, improvements, and the transformative impact Prompt2Model might usher in for the NLP and broader AI domain.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://ai-unltd.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading the AI Newsletter Today! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item></channel></rss>